var/home/core/zuul-output/0000755000175000017500000000000015154401252014524 5ustar corecorevar/home/core/zuul-output/logs/0000755000175000017500000000000015154404354015476 5ustar corecorevar/home/core/zuul-output/logs/kubelet.log.gz0000644000175000017500000177125415154404201020263 0ustar corecoreikubelet.log_o[;r)Br'o -n(!9t%Cs7}g/غIs,r.k9Gfͅ )?KYI_翪|mvſFެxۻf+ovpZjC4%_̿f\ϘקjzuQ6/㴻|]=ry+/vWŊ7 .=*EbqZnx.h{nۯSa ׋D*%(Ϗ_϶ݬvGR)$DD D~m{]iX\|U. $ॄKЗ/83Jp ώI8&xėv=E|;F}Zl8T*v (6pk**+ Le*gUWi [ӊg*ͼ~aT(;`KZ)&@i{ C2i1Gdē _%Kٻւ(Ĩ$#TLX h~lys%v6:SFA֗f΀QՇ2Kݙ$ӎ;IXN :7sL0x.`6)ɚL}ӄ]C }I4Vv@%٘e#dc0Fn 촂iHSr`岮X7̝4?qKf, # qe䧤 ss]QzH.ad!rJBi`V +|i}}THW{y|*/BP3m3A- ZPmN^iL[NrrݝE)~QGGAj^3}wy/{47[q)&c(޸0"$5ڪҾη*t:%?vEmO5tqÜ3Cyu '~qlN?}|nLFR6f8yWxYd ;K44|CK4UQviYDZh$#*)e\W$IAT;s0Gp}=9ڠedۜ+EaH#QtDV:?7#w4r_۾8ZJ%PgS!][5ߜQZ݇~- MR9z_Z;57xh|_/CWuU%v[_((G yMi@'3Pmz8~Y >hl%}Р`sMC77Aztԝp ,}Nptt%q6& ND lM;ָPZGa(X(2*91n,50/mx'})')SĔv}S%xhRe)a@r AF' ]J)ӨbqMWNjʵ2PK-guZZg !M)a(!H/?R?Q~}% ;]/ľv%T&hoP~(*טj=dߛ_SRzSa™:']*}EXɧM<@:jʨΨrPE%NT&1H>g":ͨ ҄v`tYoTq&OzcP_k(PJ'ήYXFgGہħkIM*򸆔l=q VJީ#b8&RgX2qBMoN w1ђZGd m 2P/Ɛ!" aGd;0RZ+ 9O5KiPc7CDG.b~?|ђP? -8%JNIt"`HP!]ZrͰ4j8!*(jPcǷ!)'xmv>!0[r_G{j 6JYǹ>zs;tc.mctie:x&"bR4S uV8/0%X8Ua0NET݃jYAT` &AD]Ax95mvXYs"(A+/_+*{b }@UP*5ì"M|܊W7|}N{mL=dC' =MS2[3(/hoj$=Zm Mlh>P>Qwf8*c4˥Ęk(+,«.c%_~&^%80=1Jgͤ39(&ʤdH0Ζ@.!)CGtGLS0l/LKcQ.os2% t)Eh~2p cL1%'4-1a_`[Zz㧦|k˭c ĚOρ_} Ewt3th?tvͪ{~;J0= |JUԍ;Iw}/9nh7l%>'ct Հ}a>-:(QxPyA Z UcÖgڌ:8cΗ|U1,-N9 dI [@3YN%:ò6PT:”QVay 77ĐrX(K&Y5+$wL#ɽ 4d-bbdAJ?:P>n^2] e}gjFX@&avF묇cTy^}m .Ŏ7Uֻ󂊹P-\!3^.Y9[XԦo Έ')Ji.VՕH4~)(kKC&wfm#Y~!%rpWMEWMjbn(ek~iQ)à/2,?O.Eb$BrG XF{coT;ƟlQ.Ff!bpRw@\6"yr+i37Z_j*YLfnYJ~7Z~okJX ?@?gU3U;,ד1t7lJ#wՆ;I|p"+I4ˬZcն a.1wXhxDI:;.^m9W_c{xnln"sv&{b%^AAoۺ(I#hKD:Bߩ#蘈f=9oN*.Ѓ M#JC1?t/]ps/9܎ms4gZY-07`-Id,9õ԰t#-b[uemLi_󈛥^g+!SKq<78NBx;c4<ニ)H .Pd^cR^&|El~vm9ZNi$2?$>Q IJipqc2*3g 2wbc(K.$tx2lj0<j}fa%(%#g*{C.ު s^_0Mn}#~sz1}z}V? ֝8CWA˫kLznbWo 'hG iׄ?:JmE1y<6"` T7h|ْXe|if]voɼ;ʪ~5/nQ?s dpxu\uw>=QR-Mݚίd rtc\MU&IzhBYň(&&WV#@..$ 93XMBQϷmQm{mOߐ/ۼ[𹫘qݎt6Ym|M$ 6ڟ.x4 TMXbXj-P\jА޴y$j`RO6A"EkuS"WCW\tniV K`pC?fE?~fjBwUN!N5Cu!8WqRї`m]; lau]o@ Fi:K [Av*_9B0qV(pP4W =d#t/(n&P;)_]µ!doR0`pl`n~9Fk[ٺ+4Hhao-jϸ??Rl`#/ѡZ af#rjcl ^2BQG7'Dff^f!8:/paf LnU*P(8W[U6WXZoѶ.)r$KC%(;6.qp#YI$A@EEdT+w'+'A7㢢V"+aQ33^ќz9Ӂ;=^ۭ7h9 lr_qSq-XbsK، JBJbeOfOAsg31z=ʬ-'vlQ]mKͦAc?-֋6rR)? I?ytwpC'P/9} V1}g$>8ac#sĢB\PIPfwJQH;Qxm &GBf\ZA$Ba-zk$}  L&T+̔6vmEl 05 D"wO>"J~7+0_tK%XU͍ &dtO:odtRWon%[ZihRk71VDqiގ\<:Ѓ3"gJJčE&>&EI|I˿j2ǯɘCGOrbkI%8zPU:LNTPlI&N:o&2B~ LfZ eWzRSrH ֱw;SpM8hGG&ƫEޕQf7s#bd+SDL ,FZ<1Kx&C!{P|Ռr,* ] O;*X]Eg,5,ouZm8pn;$B/,X Q"SoUG6 !ȴ ,!NB ^WKy⩽]x99?u0軡*uG"  f.s{ukkN<ƽN!n<#)u62-anOIq;6z( rx߅ eu-\q!<'J[FJ,4N:=6. +;$v6"I7%#CLTLyi{dPt/ ̿f3Nd0 ~n~YͤBOK&9<{w7U6=S0+ň+KMiw(W ]6ȧyԋ43]v_ r{i=s BG.k7%,ڃpzU4(6{\٢K 5XGU/m >6JXa5FA@ q}4B//Re&#c5t'B6Ni/~?aX9QR5'%9hb,dsPn2Y??N M<0YaXJ)?ѧ| ;&kEYhjo?BOy)O˧?GϧmI C6HJ{jckkA ~u?u'<?gdyykp<[&րj 5*<ڲhn?le\Zer֋@!C$7t yM0hFpۋ)ۏjk_vq;T0 k05Y <)G>bz?ܓvjIgY4:]U])0wMmeüze1N싰s~5$lISG*"oAEO:< iuKVM{M9$1#HR1(?b- BQ{YES-V_k Rګ $[d8'  i#i8Yls3mdHY*^Q{%w)7@ZKH8 Jt`"F'Iٺ*Dn@)y|tԞ!3Ír!S$,.:+dn̳BʺJ#SX*8ҁWח~>oFe-<uJg)(55xGn\^=.YPI 2(H'e=@W|uӀyދo\戔-b2* EM;oH$$͟@[Aѻ_߽|wY[ZNg⚶T txYΖx]eɑvťJ*V.I/d#礱ºȄWEz o\ݕj =B6wLKZ39(M +(PUՇf=sJlhdEBb uBep!*)jxKz$ޙI kюHLa^b*/H^FFIu`2a$mc Ry;R:6Dܓ>Y=z\Vee9]`3=?,};,DiwƩhWl> gw+68D& Ɓ(^ca 24߱^F[n[?irG#bYѤwd)J^;GT.Kwj;ϯfӗuaVad=$6)iI 'C~.1%YmҪXZk‹u^wG1Ƙl"oO:(x q lKxDȜOY2S3ҁ%mo(YT\3}y%F}{Ⱥi!ma0o xI0&" {4Ǩ9S`~DkM{|D<˙ik; HD;;Sٙ@qPY.$Dwiu(lO:k$Mj,&tK*iOEmH:=uݫoggޝ7[zkHKhWefӞ@蹃DPj5p=%Rwó^;n1z"8Po߽]dz+ƽnr3KO0Ao-Y*ItܴR&%b&>BL(hO7zNa>>'=+KgS{:/UooD8F?1%(@fxhk6՛tcYͼ#$ڴ7fjӞh2Hkeﯧfd|HHDŽ">#%;&,It"tLp5õB!>9'0LL]M[lvFR9I5YpVg4;RD-=w7zV\M6YnG Cyt␚ɓW%O1mmL``FiwgA`I7pWt|CX6|LGiq'N~|<>7_l>9Hh"Ki UIL`ЮF &ni4~n3?n%Kھjۅm''J[/m$g4rfrvRޝϏO'/g]fyrd-i-I6&V Gx`u.̀abĻyOڞ/1О@]; lϊ%6EC1塴ǫ3˴8TjɛUL+d6Kg^F k=^!r":#iI L5\KZ<)eM|$oB"!ڵ@%HSB2Z/"BfmB!3}2m.$%>Eü%mWO ol9%_is,ԃ)Hc`ebwUg'VKξ ,!3`˞t1R8?UD;;CY,2ڨi"R"*R2sS@AK/u5,GkңsgIz R(o(wC7:U# 2\YoU'tt!ƖۣRG5lhfd#]y=D?֨p؍H|mU4]9/[up`uV,\KCB\qGhË__mfWͽ2]Q6ë՘`ߋz/}4ly(jc 5tW6w/l/`I>|끹mQ$>N |gX ͜IH[RNOŻ !`^`Cԅ\‘;}VLGp5I#8'xal&Ȑc$ d7?rK-6xAibH)F<f7/{bv/3tOas*@J{;tH*tle[ 'y49Wͩq2>ȥ@ەP/p0Kw.mȐb#+|yɚ YZ'ދ8H KV,XH*0<>[VݯMYJXFfQ* 0kq="6m]"3`Rd1e6d逖`7ylr OO~iM )qGEByw|Ț[ՅwI1dKX]Lm!^h7YJt*CiNFK)i&G_d!gzpOb_3aIŽ6KQ S&q|pӿ^Qya5s?Ἑ^̗o:hC>xR2=ϫCy~ӡ` ^ y!k!9yL%VLU2gr26A!4vbS ]Wp+&ee *w -`J\ ppǣ}^~#_|o{ К8EW>*)D{ޛ$Oz 3,\Φ>4&$/"lX-O_vwrpP40ޢE[A͝ Z5 om2p)lbp/bjWd{R\' 礅_}\Ƕ:wNgOb{ڤ):d/8$<$qlf)n&  $uT;wD]2cM(%YjDktByxVl70~np.d1O9Þ%˧Bydgs9Q_Oʟ. /xvbIH^ni5(w>/Ol?vEWZ vnY3 cU Z'20WO|qBYAO'DvZܧdU>߆GUt\+n3X]By'zG649tV|̢T`K*"_R[كrq IH!6}_F6!ܠqK%$D)~ty(uGu() C>Gn} t]2}%1NodI_Bǂ^8]3m!'(Ֆ5Q&xo 8;'LD3} 2Syt1?Y-'O]^4b[ :,n^;S?2_~3\z=:?n# rov}e{椱I twp)U݊7it?7(=V3zD/^Bld.HSntºB4|-hJ\_"= o~#ݤ[tfv5Ytc5G/{g&'RqB(Ix_<`'skkZN-cAuOS.)L)pŰl+ a[s׭dֲcUh=Ɩ9b&2} -Of;M.~dh“5¨LIa6Pn{ɗBQiGgCXt!*0U-(pc;{*CiKe@p!E]&x!i6ٱ˭K& FCfJ9%ٕQ·BD-]R1#]TROr}S? {;Zcq6xMY 6ses~}3ɱ¹ R$' GT5%:zZ甎܋)`ŰJ38%;fHohVbK :S50exU}W`upHЍE_fNTU*q%bqD5q0mS换ohN\Uz=b ZFng&? 3 pil58IS)`Ѓ' X'f0e~v9ƟE q& z7z>%(Ȋqm|dc+lUai,|Fߟ|. d3EjZܴv]pEO7}&bXԈedKX :+Z|p8 81,w:$TiVD7ֶ].coa@>RX=4OZS׹Un %xccDa.E h :R.qɱmu$ơI8>^V Y. ,BLq~z&0>n- ,BLqfx9y>9244ANb n\"X>Y`bb*h%)(*_Gza^ sh6"BzξH( ."e)B QlvKlXt҈t9՚$ضz['.!-r"1MCĦʸ".6pE{ =CNc\ESD[T4azry !5yY~ :3] ;Y[Iȧ q>i Ǟ/"8Wxç,vܰ|X-LE7 |-[D`JLw9|fb>4Nu ߏ3j׾,$'qo%HWs\4ݞto^ X]joC!Pu&Jm l?Qac_>'"Bg<" 0^_[d0Yݎ@2!ӸvZ{Ibi/b;u8IRXAV{ύlKwŅjЙNhfG8>Vڇ RAٽ9| cbpcTfp9 6M/>_x 'ۙz7~~);qU9GLT!o 6\c_:VlnEUdn6UˇKU ;V`JUݵ>EO[)ܮCy*8¢O[cեj'x8&? ՃJȚ9![~[?$ "ssTV2i sLq>z@JM->m?" a; W;" SnsUfB7™:5GY \M뿆Xk3۾4.~n~eB|C1Yg~ؾ?jVXM w*Ǫn\7{ctMY</"˘oVq/E b\ R%.qS qY>W Rl{1>Z.|>VL h5^7eM>y̆@ x!Lh!^*,o_V684A飑i2#@+j3l૎s1@:G|gRcƈ?\H(m!LC,HI~'./p% 8$ c*Dp.cj|>z '` |]e*:q!G`G qBAgPSO}E`́JPu#]8$ 3NVz[~SvD.yΕ`|ttc‡-5 =VrPjE0Ǐ}Wd|\aD;(;Ha.]1-{K1`HbKV$n}Z+&{be@7ȱ^%u~-B竟} |6K-uT0\8"M:" /P4Qz~j3 .-8NJ|!N9/|a|>bX9T ҇#t~T)=UF"t;" 8-1I|2M+)W屎ȱL˿ύ-038F*0-)ZyT13`tTnwm* "1LӰW&jDkM~# (C>ϭQ3ߤ%EN;?P%ٱm -{2k 8Vbv"ŏݙmn&bX[cO;ӇR`RA]d+w!e rr[݊/V+@;Τ`5d0ϕ_Lع`S"cK>JG.}Ε00e>& 2ovNj31cEdA Od[=jlek#XJ|&+-T1m8NP$%s,igG\Z:h Ћ߉n!r}_R \5 6 d#=&X^-kOwĽJOG\Vf7 )16^ʁXnlwìRXYJkK}pc?Nm'`#&<6JJ4E88A1 ܎y,()i7-Ո: tC\?c%7\ct!$j´KxA|Hk6Ę_"+BLu>Ɩ=xɮ[⠋" ѫQPlp;3F$Rve6S zp!m?e8 "(\N+:Y EaSB]<^(]В|Dž8"oRs?]\McZ0ϕ!1hKs`h0O!L-w}ln2&0Ǚ'0=-T4G7! H/ͺ|@lX)+{{^s1V63 ۗI"*al NJ`Q8B\pup6_3XqCGXznL9:o pcuו8`n{ave=}OR9~yL Z)=W81Ù R$|L ]OfJlӪVVg:lDԒ'͢Zu[kWw{7st28bJ0ꨴU1|z:9lX)z2!S:'q9| 76"Q;F*04Zٚ ?V¼r_:h/AC;?_;qŽ6xMY(=ͯl~l8V0٪T zL{As:&EXAL7 8UgK'lT)*7n`R|=(B X1|wr߾6I ux1^_V2޲jmi^\b``Q3dBxV3NBk1;H{6ά/KeFJ*n7ئYbQP"eDU]dJrbIx֍ﶾ$h873Mnp<&Dj8m cUM;`و5H.R#lT(AJd>^Dƪ0hV~ XmVcUi>g~jPhyU?V]xDcUR2QeRSicS1VED!Hh0ZŨY6eUoq{\~*ϣB6Id Sx !3?Z5H2A$Gk)Xb7yF_X:*o$M`M-"eu,7Aatt9vX_ϫAO= +k|x8JgcXӢZ^")$zg}Z$^BjH2~SR퀧NI n8Q;ύ/ ȲWo[}|!~^\ˋOxϋ+j[n,<䍐Y~!pcz^\Efl[A|`mqZG" 5ԡ^ݘE+MZ]@NlBxW,ke=M9"l{EH?,yߗ/j]PQAțl<oxr˜(thǓiYd"|FǮnH%s lv3-(sO!n/L:d'礧),B8sׇLMa.P-~`ad>e$w"clѽ4:F]o 9Qz?6mѴ?G~4:`S4fviT֣j?b@XI4~ӡwmj;@7,KG"kܟ=7#+ݦi{#O|&*t_^ bK"aDl,?f+ J \d2R-dnpWٗv7Ē5u;%I\hlG T}2r۵ꉥ"T"$YT%^^&*bZpmvwVIYQE)FϬqHQ|\)SPxTF,Iނ̫-|~"z"> 5eaܲfI& B&KQ(D>X-I5)X2PrR-eDtI5 וՇw꨷_,\ɲrUUլ}J,Tu뇽$$g1IrjN-'KGʚlCJu)$;̒|`6ުd\OT~tWJ J&qS7~Q^_Y4 @*($G?b@b/(ن'[s-#tr;RyT Ŋ<דqtFD^R!=3MM+r0]ȗe39L9NQ^N#4giOy-bVr-gi Wzv{ܹ陷ł)m'xtI+ϼ[cra^,E)ΛN>=!+ rr//Y^3`C@k:7'[P&l놥E$gq]6jy;n-窤r!\DM 2{:;JQw^egv^~9U޾:`˪aCudmg|e};_P׵EE\(H0,M1@\֓kO/NRU5Fc @4(  g(oT?wM͗MY. ?yJC‹$8IxlpX&LW8 UT#k9|Hr_ GժI-{/ 1N#M%I j (Q)eSsVTK9-2׬foX[&nqtv{<z[ iV> @ gzuCy7L=@4IB΃Q0\W& Z_'U0FE@PɯoO޼\IȜ;BQsuT^ n4ybFu Go2T x|gi(E2.6"uqH,,)A:փȍ+.<>YZ!EY{*ТٙY&#q;>:IM}#nѧ,V- 쏦qp R>#YYQhVavu\f3 AqJUï1oΈ$QGgw ,OJ@mGh/fQ2FjGv"ZS8>‘YƎTeo1>jwla&` UN1\1JZq6Ͼ>NOUaٖZErĉmNU8^޵9>.qJՐm_!-SnG`6YK"W6Z!0P j aX2Li.\[y%C?k#$z-g'!mX0rȭ*Hoղ ٺ,RTp6e/#ۻPTIJg7:R ,`Cq&wEkXT nj/JǪLo:eI!-T%)Q,wleu# շ˳g>ؽnjuR@LF7[p|-mcnZi#YR.tRSyaW,IU";PXa(٢1( nOP{J9_(mĥԈ\6\:/3ĥC+B5W._nkYREb ˄/OX瘗U0e{.#kfl-d:BUo YZ7tV_Sc|U-4dcD9nJ(< `j9ˎ"2biKXKB Dg.o j..% n|WbQ mȢm!X ܑ`OMk lF,PZ nC"Y&nJ0j>XMbi[w٘vZYW(nxV "V(oi֭G0\oK#0֛`FSC+JM=8mf,@*$@OTnx,e6pvR<2(gOgn)ͳgyn]sEXGѨs8V6}qk;yT`"]-TFS"$8Q$Um4ebn/}B<]굏RHC;V%' SPD}U+@BhOno(2o{nmݾ Qhlvw4bQlkZ×t96uM]gڭj+MiSѶ|RbzmǾ1uz|z;}F}" ?.DPwұVMZ>I|*=I)8fhn:(efT`PA#j0J8t}|a7nc\dT+wjL+&7\*lq<Ǐy80Pq`njZn:Ny&\讋(vo7؅ ք\xJt POG#i -^F-`+@#ݟd5P:2Oo6"$)yMQh-*Z ޤ!<]]и}_&sWK#->uh*u=''UW wC iuC8<}驫K݆܍nwqxMͲeAhjVg0iiwm8 HCt^Ȧ'ҙJD̊u`1ufdO&M(Kg";eu.{xx|ޣ@DX8>j?_OqtWd(XU]Iu UX{Ju.&}qzf#T;+m18kVC*L4vtbɣ w7k}\bݛ6hm^b)2Q;":4ҊPDGѡƌJC"J(b_>w0 o'ϕ%iؠ1u@Ϣ2xa$F؃v%/Q@f1le&Zz~Xk t"NO 6_mLw9wÚ2F|DYH 224k̮ g1 t)}eͺLa?0jX LGOΛ@" W_a?>' $(@:@9V><}=WO g}0H07`3K]v ,Sgc &bِ%i}( RĶ3|㤦y[1q5$j&3H7|?{۸ !\ܳvHzA&mf7'ٴ.YumW;CR6HF=h67ppT߰[+/;Z0_LϷ4 l6-z}F-TVj3 V[9Ŋa-X,I3a}6PZ-bxO'dg~{ws$>g2"jG3uϒIq'm [Aj F> ̮X|O<{}ggpqv>f}ݎ|.{n0.髡y:&_n ubE5ʉ% yvU/$w&)d*p>~Q&S'T`\_Թx cӷU\A*#?~_L2آ>b" p)Lkً:7_8 Fd`am|>EO\EfU1QRb?Sj|LEaLDxe<ܰ6ѭ>5cLkt~(G k{Çyr?g_~pӓȿn6'A# 8ҹ:hcix(}y-͛X]N`aASw F1^XJ>2^ (lN^jBO-+|ԣf<5a"^m ,-TA;OqP/t xC gĂ"BV}w:JFZĥ藔sߩY>l-V'L\!'G`N4$Ozxo&~/I R?^eMm۸={( ꊒ E;d9Q0lP! (lF]d*jquq"<{YŔ  {x:(r?BՒpܦn&#Vfєrzpa]0Lou孵BţkE\XXŝuLE>q3Y`F$#Q^}9vb(w=(lC Ѿ{Q/$+`Ly?msNz7!"'$'3v]| ^gɰ{ap0v(_:-oZ]%*?EdW)Qz/]w[ }8,B 74is9X27{b^#fnH(fx^x{zQo P`Ȇ1Ѡ4G9 DNW3,t|րvD:y6ch`cR+4U8\(k뵄,,V&̺BG<pS"?pxX$ k ̍zgB .+s}?#n/ ]XÒ O1Fcq滏T23>R!0 k<?I2sGGV1,uDb&Aux' l#&Yܠw8 #2ay z#2tjfmЬ'bX> yE䯾ͶƏ[X`Y=욭pTEE`37lzJ4E۪{3g^8?)܀elyG B;qtws~bl}\\mQ=qC~I жY:x3‹Ҁez~!GU<Կ v.({쉂]]kRK'4*{f6+۩]v}J>mܸ._ ,]Vխ^2—r]]NK8OcՃMQ05-f|FҫzԊ!oa0:,&IRBGhP7ϟA[`ZuN~.&;pN *(gza7ԟNj+ƅ3+6*q&U&S'TM2Çj}yxi?qϋ:iUHqlІ"m9VjNyz^\po,1 P&y +.u9H~3tTrl׆ u2$Do6RL,hhzPUp(A9)&gy:Jz7y 3-l~ӂˉ+ @a80/@ғjJlij~LP77I L?S9 6J gKaja^Ȍ#}z齂L'z`ji?f6A#1]4yy11w2l{ㆰz{M_j>Ӗ:q#LWn2 \bqFC x2UX vZt]Ꚙ2yB2;d !jB,KzPAj>?**-+N,N 䳄9D$z= \nQ\xR WF A䫞%SkUhQ3P\x "53ymmkZ@q0CcG"'!(GӜVW,0T^x'TBk,<VV3߰ku~MfHm.JYJ^T]pQ \J_Ȏ; r:DPOpBX*|8@ogyU>[ e|սvҘ z&"7FwL~gd'.,^Gyant ^:=J)Rkk躬g=iy[A2N2!MFyZYk: ;8n?+o'Ia0*#{5ajy~tfm^g˻2H_m+ҍ&E!V|2>p6Ri J`^*%mS#沗q5{'CIMD2u{ߙOx4ZW3XɅg~ϴm"zʅ)o'4jzV"ljP(yd*QM)`/'Ĵ,+X)chrRdۓ+S =S\(CoѾ$uYMf蛜yf<^ jyQ֘Z, dvxKظQR<1 c2w񫴲 NZFm˷ ]Gʬk0̆l{C5Uohzj=F1dƚ'׊ŦVbf8WC$gGy]ko[Ir+ :cS9"-O%) <4`Ȳl=wZN@@ڕ]Cx&&2_:v>:1!Sݱl5aH'~oWem~ ]⾼ [lWm>k%oxpX/>Iy?Ww}uթ֧' M׾LˡD >^\Ms 8.'vwu~bW>hg,~=vD/`abIڹjyǟG/7<|3)>!o>{'8ej~OZ6zнOP V 9Vk> !d ޴`^^߬_Aԁ)-؀/bٵ=./~ǃ.Y=mY-99x@oy ֟t?mCht@L|h9^ECH:VR)mlZlY]/W).'Γ~jBZ%̂͢ŦW+h*Hb<"ڱ긺7%F\3m`>kF( SCl;?)MޗS`I μM+H'#șώGǞA[u>&ARcuNk9c$&?,[ZD9دrLkxxu7%!q#0>Fg^!@-c:?6bQ8dԇVxV#ttu7?S5:j@4GkxW^:6uIdqg̽ꮗ!a+'z&h3JYʠZg~L'j>Gj)VRLK ڬB.BX81p۲}&qp1B ^ԆEWU)ru2N/b$EmÊ5+ Tvsjl}a,H!MlF$tF|aM~H-{"Q"OF cj(pev,g@Q3KQe YBH:Ry@# W!sVxDs){mሑtC!Qɳ*w=CTkRlOMfTQm%>JT  vadҔƬs-|eV҅AH"4(fAQ9d4h~Bق,"$rjǭCk;pA v+qDъd 8#tdDJ5uyVO? nn,nZ[ U.VD+C0`16愶L!6eeś.bpHu %\nJN h\iZGLP4n‚ǯ8uI!(<dd3w0b=&ڰ'*q$ArpT$_lG%ṈIbz7.dQ}_Vӌ(D>BD5m[2P@'<$E%]’XcK&E":%jgX>FҡsOJvg Fڳ0!Bٕӂq&8w.DHP)3`N0ff9^k>OzV{+N1@{H@TܧkdA&LLE iø21쩦oۯٛ&݉sa{fՇ٢jbZ Jxp t =KRb-#SIX{IZQ딘:eKk1ʍJecİ=14[/= {nK/a*Ȫi#5KF'0jLBD.:Eh> z1ɕ=3rdOfCH HA%lہܽpKzQ.3xmϖG{^|ϯ8x۞ ~lK(P\ZWlL[YM\Pd=G.bތUD8Q,K"ZMH S5IdzGg+X\άop3/ oH6=#͓45 Q$HqVT d (*qJSD 1h~V:32 Q$8fT9@h7\\@9CD t8p:S1"0<ʀ> p̆N~&o,.#ç:l6AyHiBN%de@Ϫ<+cG{zы>\A$-OIY r:#kHpW^jl+u) 6UdO(8tUco:5Ĉ!b ^RVɴg4h]`Y$8J}}!0NW:P&^N6g՚bFbº `$陷e煔VJ5րx<0KaXv6& 8Ŷk0k'ym7 6@z"/dᨄUgL +r6ǀl3H"Vjpdh>?|Zo{Bc,,ˆLTgѸɵy&Ú_tv\l$sRKs]swd)AYL ڇуEB4;OS/ }sPנ ҃H:ҀL`_TzDq>2F4'N_#: GCT4}:+^~B /~2ZՐ| bΑM67 >j l$T3x(ܸ蘗s' &8!$_GUR«f}|9+X҂3/B ـ4Q\ `=4FcvFo=}[?@7j-P.!5\<6mQ@wjpgrpQ-v3PY2^1eT9&(™mҢmpep9q =`)Yjb̶6C88wqJSDz+mƶ `Als_ A4^a'ɇ`qiũ eddwKrin[ y܅ G0 kk~? {6qږq݆gnЈ=߀U >؄p'82֩%Z:K;_cTNP47^RۣFپnzLԔ:#0/v}T,պNۮ- :sŃA ZuV;g ]M_[hl5Z,ŗ6Ĥ2q&oFG~S2pIxl^15+ŔrTxRpop `b|IYU9뽃Wa*?`5Vr/p Hǹv38BY$~ÔGzo/oY3!|7:ϊ/s$q#~#226i qDBy6#c3{wY DYǵKAXbIÇX1qUZ&IT?؅13|S6N7.hJ#QS@bN?lŃԏu&\>6txLuq޿1yRd?^{;z{\[?ɝ\xC鏯NGt?aOII 凍ZXsB;wUUMI68 &~:Mt'?Uo4d]7I_-<I}Cy*o'})6 qMr3;M̓dlo= F+?'%ԓbw! K51W2?M<%y>ʧZ{'3׏v `~9Ut!JvWV*?1}4)9AwA;'\w͌*Q3hxݒ^l`L+∂,&.z $ϣ7T! Ju)S!Dz%C |:Z XX\;BMX|޸ę((߫@8 @@lW_Ya97ѕ/[]GUU&% *U IQlwaʩ(77 6uvKнJ{( $s0z/g83s+/CNͬ4-3-1k3>ܚ'S;(f; Gj{ΧA;E{wU(\\}Lz3lTM5͐#1T]oNqܿWfŔ"?QrmtIN.بZU)6V32TNȗaqh1MM3ňsoEI'`2+ϨC"Rr+59ZlnykU1p YQ* s/E\*N<i$_ĮɊ??${[7Aj7\n6@YG֒[D0NP!cXGMKYGg0*brk~VMrq8S^hoSX} HV6>t|!Z-]S`DiZJapʬVJG,)VU<: ʰI1o[#Հ%lcM>'ry6&[/}._&<$,"pn{=fm+|ͮjy s(ZsdY8/ζW+iNk.fRÕ<0R)*Sn~H-^J3ݚ~B8؎}y_.5»uiK#Ѽ~X%18?TyJס|>o7`HPCM =bпHǐmCiZDŗdGuD87(j q+"'QˍnNv6=vr-lȗ^A^RVeGi+M/}PtW.M.5xWlmrGjZ8K/?Э=~v}%ǁ^NKaD^Hj}D I֡:xҙ:(.e] Ԋ:xe_JCsF*k)_|Dz6G+xEhŊbENyuY~RcEiy0Mĕ}E_w7%V ;.(/1ʛ~ұ/J05HJ̦D[K{!γ(eaq _w:.iNB6tt6a8G NKu@o3.4ky`КL(PrȔPDEXSH!ElήM;jM;f0E{4cUJpȃ–;͵Z%u . BiSM;jM;ˬ5'6<~uyqBp{lj} F\R# QP!"24ќRM|% ܴFiO ʷ65CjȥHB6)(s'הv^ XښvH3vdxU Ĉ3ufCھl$-|;L5=R:n'`?+iJ)IЎkO|C~ q4E⻐{+Ίom#ohgl|0VJ1\ݵV`N {FU콀qphnGáp{ Dڝ4ƌߛ)o<мjv`ZDEஅ]`^)}l)Ǣx:UHg;nhˠt~lN21|mf`/" } )3~yy*ƕku հ1VQc,i*Ef"vs*s*xC6TZ9ogge5R 0XeZ!C8& ^`#O  t_\o$)jC%XtBե`7p%)锋ꒋNpG5VneI"` wv`1`4H$aZpOFzKYxK|ƭ-^$ƈ5! -V_zx]6Xw(q>pÔ( [J+:"֔mkJ/Av֣NQ1)q͌B vfTPK' BƥJ:GC`:P EαDL[pFtbŲ+&2~6 mJ6)D4!iD62C2s&W:4`sҀА2e)(aV`SR֤$ zH L[AΥxy f F Y aΆ@O昰F@jHB){5GZw'$zn\( /OYhgnÇ3*QLcx8B`pp06/EVr@҆ImK LI51,!Yc4d)%QuC>sHWjg`5O6CKTRvsr4O .JtDxޕqd% #RAu<`3q`FAg$rHʉWūyE5݀mdwW]E# &(S1,8E1bNi.6hxML58mTLi]H uY0u>榭s+0GDkZ9T-!f* \5J+cDw"/DYgI"E*J`L !aՇTҥ_D.Z:)WTDJLƌ!86ZY0FEQ ]###TXR237< ߸ ,+ߝ?];8Н~-N\L7>(`,_.p=&cIV&$,x}YսVsѢp" fK7y/^ ?0Mୟ,|~WOQ„zj xRϱ&w״LeVn^(aUY0Ic6R` X +V7$Zۗ9#B9WPUcLZ0ZUtApU,P-B_(𭟝UE`W~b)'YoJfu{>^[n\U7o:KobLhsd~.qs0MMkݦ/"V 7䥈H`6R.Ykģ' ?v5PZkPF F90[qF=0KN{TSJ]yFB`­8#K5B@4#3f4T:\eՊJL@U5.bEԺ(Bd13D}SC`Zf*?\ Ug"­,Fhh 9U|b7hҽTwͦ`C֙{t}t%YByŵ-L .<^N?A : 5Ѥ1MWUJߍnC8qJu'][=ÁO3:wl>[V}99d>OEj#ϛ&i꣋8"J858ɕ: }v/jőj56~Ss@іjvk5+dzPf6}|E9H/nO3#rure#Z4RVd? oϳƿfh\怈6r9B;?~!ܟΟC^]ߛh_.|9c;x)>m:p vwdZ7: xm4}o }蓪h:s,_vGx~zx3"vWvG3L]v7j!fn^E \ru}F8s dDMt}F]3XWd@+vue'ö$n >ތ2h`hV_\wrnۦs Wm2,P+V#^%%sEE2Nw2oY;_f#'+/gـݟٯW\8JvF|"fo? 4X@|%]t~Mؿ.4,7b(Н_]Of\˗CʧNOߎhP'd@. O9Wm(n\zn(YUKw^\׹(qf_WD,Ͻ_"&NNz{=sڛ{ϯۻׇ6\fɇ^vs֧v4'IlЎx>[H+;ݷ`,NJ L>k_aa8".jE(T&M>я~̖2ku"YX@š5f,_B3IގW#^FLTdI,k IJ` 5-8(OyTqX`[2|**S3H``7 ՖǒŽ(.J"ц%VѓEEZhרP-X>~Bx6 Lɗ%؂ИқUP3Nra˾T-/'` 'k{.+hb#ΐIs[0t`7qtt V\*= B-dUՕB`J?Sl^C0u9$DZі#V:> cǼҙou&F.yḤ^ࢻ_Ke>8\:a?Qoy>;c7Ĩo]RGx=Cߧ^<$_^t%bu/JXeDm{fK:j:%zծΛQQzjYϹpUA[30EzaF`jS]y0&Ap A+]++C)Q)O_.zwe7S²r փQT += XJUզU6ޏ&jBԵqV4z6vim/0(9sd*r 2,hWISG2+B` yfE(߬Z"33OìH.fE ̊[рV\uH M L+'HzPIozu/%h" iv+ W\ղWZJ$)t݌,}_YPlAixGEahgKoȍE2la;#gU\O%T=C6/_ܼN&h:?K͚`{K r~8鯟tf,j[D۽պ7(]n=b]J4~*Ev* ƅ?S=ʓB]0fï}}M/N o£T%֢$ dWSjy p#om6.qud}WTB5N,^tb%,deIOd̢A6bMXJʒVT}t 1'dI)(&(&K`JNPw"2;8>R[ּ&S*s2vtpD8k n^wpx=95 ژޫvsbI ˮ8smpl.EMAۢ9ѬGE;d܎yAmjeiſ..#CD&m/+ f^,vAY0MnJ%X aIe |m,kMb qqQXm-nsAŽ).Ǝw߆GwqQ*b)i)NjV!X1W.!0wĒdseŽ҉p'E檙R[&Qh)E; F>CA_u)[r/`3ͩ44XMNq9oӬZ}z+69l*lZi%]<ʅ9=Kbݞ۰_VmU mׄϻ+s8.{i|ze}ߓˈ"ƍ@ pR!5q#XTq;N<,õoVt|xK \c*#I)}l5/#J Ҕmd4i T&[y랛6yn}'"D؞z +ekn\5KjBXٔ49*&uq WhiŵlVBJ5"\ɧ(v;lWOĬZ)H)͌*BlRJf!s#5iTb!X3P6M])sl[a4L,@{87uyձM 5F*:։2ZK YTyL&DF cdֹ:FHi&QRť m?ʔPbHow|cSy{t߯r$ܝ۟^PwfI+D=@+2# _cX`_l4lE-Iu"%,R<.fp=qPi?# /EEEj$_Ne6'RGQfHUQd<؄ ,_@:@j`H/ a|Q&P/m:$^dP!xiFkM]&e>5vh9jST?4(fU-.7B:te&S1~~6m@G+"5% rFJj-N?ڴAn'?MRJ]Foh;gm~֚)S Yݠsh#t S !lZҦU|)FO?a 2zɴS1L8H!^ by`9B@S&rFqUHo@84ۋiRQie,馘&k:a*,59d]Hoءs{nGb!FoЍņCJ 5o57Jaa)m#9B!0mk5cl& g5]HoI1B4 $@)Kl0Z!A d6+!5d}C QҔgDn3|kʢ*RJ6JJCEb<, #y41CF2~2z*댉6UͲ}sZΙ*m\9[*@W9ʙU|&5ǟKY:Ò]_UR`Q ZS*{Ej15Mj&;Ej(۬=pJ=Z 7@6|S3Z(mMA(d=cx2tuαx[B !G$QQYF{e7*k9) F@C 2rA"4 ]?:Y)US2`І 4GGhlFCgBHf{vBrr*7RH?YAUQhOvdުmn`PZ5BNG@$}yΛvYـPZ VpK[֢7˨,\u=I2TXm'H  b|:zw~tIh~|x{j^Xoʻt1!B{kᔷ'^,pʻ A&@qޗg$pc=s0YǮ1?eVln[v]V8hA(utۄra@DmV'[7ISYn&Ӭ5h!iXP td")htb>"M1?dJPx 22d7e6(AB(u9LB*St˨ʭOaQu,TfY(@sN EހA ')/7S"(~)@ΡvH.B bsUP\<=).ʻ~<VH5Ф m5 T йZGr)"s;1!b } ij&x' :,Bk,:s ,XԳUluζU{Fc׹*!T:KFt ۰ܭasPSJ JfUڒSiT}3fJ?mݒ|+Je:*Hat\UF`eڬ"%}b>dYæ; +F ۸\M?^5QYauhu@ !QhGXBjmT- \ߙB[MJ؆.JDr&DۅF%hg9[Y ?)hoHn@ /BAUHBʝoYᨊP`Cls \ W0&$2 7;Hm7;U9FɟؔBzZ ۧϧ\\jg[;Ϭ "B6m6Nn~4il櫝d$|5敖PZ p*`u?[{ޏoGNYd2˽/jC9$7 ѡ85Z^I]ޔU ٠`g▇2(˲7#^7>ތnAx%* }e_;>IqІ__sM"Snt$dNx?#JwKJ<|3 >FZoLJU}5Β!ع/ix^zn UzH?@ 8@Ń@AQq=P0ZY,]٥{`NOji=ZϨ8|;F7$7asRβ@˃[[ ]D~lGS?ҡ _W^4e2>xA|:zw~t x@䳚2 My"ǟE9nGbwﯼ;~m:^p: rJ}ɣ5% YR 9ڌanqڟ~Wkϙ_y43܏8]ng_j yJ+ X3 dv1.Iu}=B>B=)oo|,.w}|2xNvy ]:ٺc}o&tنk-.x|?tLob'Q!W,蝧ػP!ksrVϣsP/I\ޥO5=%R3&ͥ/o"چoxRE5Tjs/,KE-PU/'0Bih[f+Br~AťeRѪ<[4j >9cxH'W|q&kJ*@2is-LYZdEAUf*RUi9iKr$O]\u|7Ety~3ƬvCn6޺ז`%oaО#uYNvgF&^pHx}T5SIM!684]L] }LNp9Spfw [<1z_/[^>ևB=O&\Gև1O>N6y8<Phao}gJPX5|A]>`)t|Zma:.\EYLFzy//eyzwQH]!Ntm6[R #9+XI~ 0OA ~zu^p{RPiڑvi+JZj_:S;Y1"c.3}Fƛ%T˕?M u*qYNHYMo:`P[PZQOv —i|c5>K?Z!ڿ 6lAlCcJiʍ_o|~^5UOs9^v[isu/N/`)ΡG+?u] J?w԰7||5bz˅+vAH$mw -U)xg?|{799ԋ•OO?֊'{ NyuApͬ4' E_nPiWT9ֳ_m 1]a$7o|b6"R$sb9,/) LɋM)%1/Up)fKA%FD@yg݈hY,w#RnD4xqg. [ɲ~V{>9?O]lWt:9\ ◴5 0$] ŰI+o6n Lv| 6ҸvWD+4Wf[$+ϣ²ѧyU\%ǝ4-Sܾ嘫S]='ׯJӈsn$SjF7\M w#6rdՀմixdIH}X"1oF%e-@fn*FRBۖ֜mpx;J+`^.A1䗟 $ۏ$u02l|͡PӾ1X_k ~mAbE39mj)Ӧ?bӴ 4n\ԂVނU7:Mc^&ܫfen#E.b J)Tc;S)d˃R&n[vLoKI~Wt~3&Oƹ,UWVE i$;;X ؐJP ``UEb իi{@ˠɑu`jhtF^hٸ< urj{a)('H׫zg>Nm& nuA}SqUbr2?3?OdXGr/޷SUX-Vpy,3Dd^Ny p%h+ q ؙ7-v^+tQFm^s]ؼDСP_R?Hy1W|I/fx7٦S29y ->륬c mϮXukpA2c1Uڗ>%W%F }@W.?Ԯ ,^v]3}X&$jMb ؏cgfl/vf,YeV`*;~h#`5r[Ffs:^nY"6D~~3f| Hb`#Z>)Uxo56KZ܉ MÎ0mi = iM0dG?_ ccuvZ?_glPGeo~-jt ڕ }"muZei սSĞ65HHDΐk!DSjVf?_TL_o^wҙ)z1wKA))%VBДi\@ifӛ8OQcՠlrY@h6J0u=̘҇~|=ܰ-z23\>)\b2Ǩs"wQЬ>nlr5ȿV5>g'E]?& .nz]Ͱy OͿ]N#CL&M:VTݻp|:;ȋ0}d\VL[qgIȶ3)qZhAs%IIgb6e s;{cE[u-05x684=S>fbءc#4 (X3 (),ZI8kJ13 k)VX2J R h/@bʍpiMh9U5Mg[OVA_sf*Zg05pw[H `j+ ؍ q(:;P-PrN9Sy揧7OKD5v =đ OܜxhNb_3A:-.-Grw}k؇$P2<o.o|am/]Pz0Q,t=\u"8tZ};"y=0g?dŐn).~EߋK&ݽfxs!/$ ̢"_WV}7?uTԓ͏۟ud}Y-b6O54CT9fQQGwG8ty3]B;wn#Vӿ>Xh-l]qu_K߿_(E}uй|C;Z*鮪Z۽[Rwu`o{ݛgg_W{7xuN|BӚ_^} my$鏧sxw+O?v1NZk}m yW`J/od]~=s x)G3O|6/?0x'zZLajdS0/*\|Z0{ Sq`|p .{Aʖ̵+D:8ҒBUs5f˿t$IZh=po]_-~`xPӃӋK^yd񢓣lK0MX8)I\9'Ң~q"@fɒ CG)O2>$bH\8Òq?ƳvT=.T*Rǔ9_ՙyu#QID +Wچ@.,X{bL$n\ Ya9"s1Jq9J O,+Zǃpg@$u}%qU ]tvCw+N"p #„ Z-$v  ZBنqϋ `b&@E*,f؜:@iz"8j5Ӻ~١[]J1@/YbrpNF&*ˆJ0j`MuNyXߧ);À-?,m-c%+(RYo,` JyTLÇ,vi`K߭; S]V: .z-U:Xu򘄐j8; `փU,BК<4D2YHAPPzdkmƲE0 |\ 41;I$_|Bے!ɝ4{X%liAXS}[Ë9:l{2 eKineAq 2q4z,:sP& :˨ϣVweqfQM uY#aK+U}smt#,d+cCX,𢡊Xk=xhHc#.{Ҁ^A4._[0#D LBQXYBh D;1,, W*vR%m֩rFf,,0ZnJ iYGP<2 g4`mLAd&y,4<(?= D-![(o4*"J$p9 XTYxa| 09,+ )PND$EfBIQY|A AGÙ+̠j$)8B*+EA(pU_%Cad jn%LL: *KT_%jJ^%cŐJ9Zỹz'[t>踜gtxMgysuLp$mݹ(!.n(!f330*#`=}"۳,fw2y(ѣ8)xNWrq1PfKg8(A/C]H樇"xG9\6N,}Vpwz)aID@7BԲ(p))dRj1"j<XzH\c#dԆaVҨ,RFxn`4HYeR `r;l'2bF" 0AkO2xGVK<gP"?/TR %q` 6B Fx,0¢FX&PZ(?pQ!VzQ iDF: |G qɄSZ(T uJHC0$nz%7JqF*=cCUԺ [1Dd ¬T"ĈPCJ,y褀K[ YD**t5շ+!8m5cUfԫ.~=+Z0Kq 2&Tl4`N1*|(nvKΆ~#fe E&pS WcdHbBw^C%h%"[wvuڷwVg=` oyiφ~թ"ŵ5`׋ ~}#}_JzfnuyĥAFuU|ffz;yԇ=ot[Osww g\n\3/~PsZve=S*L?ng{[Lզ行$)ZC!P+e3z(U례ҨPuсzzzzzzzzzzzzzzzzzzzzzzzz/[LKz( C.w血례RP32롺롺롺롺롺롺롺롺롺롺롺롺롺롺롺롺롺롺롺롺롺롺롺롾h=* |W;eOtյ<=)VlicUZۊl )k˹j/]WQ*"/Q'e,a)#59q="|_b|$sN˵1ς+ţ85bz_6e >[zr56˰sNJo\Djm:./a07xpXC{RN9~y~3O{g?H~.`Vxa +EPzY+UD.8o.s|sgf>3\97^*0ʖtT;:@ƚ:8)" ow+9R<-UesYڠP²P'U (ȌyF0w]Ff6Qpn㘯&7ӛGRN wIk!xaNW~y_}|m)ˀOL0,@rݥ^=| 'X`cQ? '=;Ǘ:鏝']`oCyc kIoȸۖMu=bS0ɳS߇w!Q.?$[_+X#5l#`4m#`"~\bOI\:vdKj,!F1zy!`1 i&ft+F4vXdH8ǎ V VX E#`9)NE X (m#`sZYFZ?x`KP2~vB۝b}x7bO&Wu/4dJ<:{p-Ց^(<$BIxPxǝ}(xf5_(4BT:cYj\cY +#?",)r3PJfL+eT+f`hkZY,S}RΆE\9u#`Rmedf{Xű`ΪF-n1`zOjײ\XaR[%[K"-4Vikq05cp/~P{8j3BqG ֜1#iy+#+ + #?,D#`UݫG~M?XcF:˘==3f;2E,qo[_VFtJ7V-ZI UFYg>ruΆ &+cGV&V:f~+rBVL; ,oqY8?X|9͎^0z*0uOop3t= t(t#`Ɂж` FVie6tVyd`c\(Xθ>' _ cZYb5hxZ*5q |5uzuq4wgyJۤ].GV >HZsDS?Z|U} HKFV2 9Kꓯd&\1}u=oub/W~ ٯ~^׃)\p6p||uI|N64O6bW_6ZO~仫_ MݧGoz3h&_TrX˖놰D󼊃zp~/0w|6W-4)~оP:D&`d [enWS 2ڜ Ki*,rs,ɐ"&qn#9S,ީkQZ+TC(՘-T{hSJ \&!2'+'aZSÖ3cK k, ^(JIahM"֕g$0f飞i4rkx ˡGQ#αT=OmV5\qeGYxH!ّ}(I%^"L&.7ȗјjFYE"A;S^11% (x@<0F3p-琺WN<845 &JBN ,tЩ%.fdp/C$@i÷ NFpā4Tƅ[LyNDIϐBuF\FFU`$ q SI6 S*!0PmXEں֤|N`PMN+|"vy=hǂqBVbQI2a gʼnQ2"uDh4H`娠($T)S `$(_TřG2˥^f@~ N:֩R57]6ǐ?!g0c)(hU!ArK @ iR\4Exdw=$?@Ǎw4^KnYy4_Y⹠ QEV}nVQ@UEJ2OFǹqʄ5%@A">̭80h]v9shB huUnm¸!3ȂgJ- .>Lp&#%RI9n+XT507j7ժ%h~CBYG]PSP:P]لT[CQ VudE>\,8-!)x^>WujPHN:ZVh`8ܛuYVI` $k۝@ArͨZQ |[`4q;A33r@FE~bt@rR|\1FCu"3;BL>DDߧia; &~qoxqWnjʖ ;Z.D n|@ E bdr7)pĥq@8mXYd*@rr2fa1%Hv9mk/=*X va-7&%EPi$Qd"+3;p(Ե`8Ӆ`yOE#"AraRVex$ ې, S`QչQ,TG֗>^[Qt0ۊjQլa$S`b~ ?^/'zwy2W)o1%}6Fx4a=\LۛAʁ^Z@mD%\uj /o#dKo(},CJ5KrN*G!;Gd@z l l+ lwkp&m>Q9H3ZMV5-A̓_x(8CȒ,L\3vy@0QLD4"(?=D {(o,*"m,jXTҏJüOD(́*ΒM ʉـJ6 L'?!ud19MUVPěAiPKэ z+"*©,eLU[a-WYBbȔZ 66Ѧ`AY nnyqv\j~<]/&ZLQp@ԍ7P7ӌTGB`:SsGAŬMŰ,]kdly$8FDf cdF@c t 08`mPi#b .`+=ae J ΀|" =(ޟ&pJnÄ@[k|T0@֞ JOPX2T*`&D W .* u{cGXb *5=<.!`>옄CU`Fq*ЅK0p%Gx !K#ָrT#Lj~`!(t!"`B=`j?ҭzB菗m~7Rr\86wbrL> 7wKξ8{Uuj{1h:p5H Vg,j/Ak5.ou5\/>W^o= oxBb^ou9 R\;[8o\/CAs[c+yRp5f+9Y/V<J-lLv'[Oϧ=%t\^-޿{Gҝ|{nol $|r|5{x{^pyYr / ==m54=?=s{{t=oHex9aCu=TCu=TCu=TCu=TCu=TCu=TCu=TCu=TCu=TCu=TCu=TCu=TCu=TCu=TCu=TCu=TCu=TCu=TCu=TCu=TCu=TCu=TC}P]| $Աa:BR:7.M uB.B.B.B.B.B.B.B.B.B.B.B.B.B.B.B.B.B.B.B.B.B.B.yBoҌq:t:2JQ(BߢP'{PtNtNtNtNtNtNtNtNtNtNtNtNtNtNtNtNtNtNtNtNtNtNtΛ<^ٕNqv%{Vk0ɵjY<\._t̠X-S]u6Lr]/fg d ;bl@4XU_Iux=! Ᏻ1j~uEC~?~zB/dnOwg`zE>>%Yqs&bAXGu/#Xy"&VLA5JDj"D"=h` X^ 8EM؀2;U;U'9pR,AO$|`?>E?a%(r :T50Zt=={5"vO=Lv*V N!^:S5)&6H^J~PzGDH ~"`I*q`w$8X->irk.v뤓FN^˩8Xm̎`͉M*D/PJE~2#xX SY ^ʐXg:30K` LwK*8'VN٩Ldvk4جRuZATla֝mK/&VjsnǷvnxyK@/o4y`c_+uؗ;_U]?[گX - -.POZk$T^zZR"U.ѹEFcW _oOjU@X|<8~z'G?򎸛;]V;"ۂ"^r|<}ϯ=O?񼵙}qj n/ӗ`οI?~M'8Xe4(XI5ƒuWu &-&6`=6~~;XB08J+M ׊f"`UO  Ukj*#,ǮTF6I%2$vKm'lALt_4 d&UX'~fk=wiXC<|n)u's{ N}pV6D'iD3jLTFV92gٰ/-\ _2[r9-n..a ţ|cw S !b&]wpWǯo#iO+i?FuhoY}GY:CdR͑6'/Ћޯze}_ }G=٣D>aǮ}|H̆G8Z2DosU\R>W-ׇ 6F5nƠwWЋ}nʰ3wtyK^zwn7`B\lWx~@c jímQR mZ*-W;\+{:/y kbrnvkS=[,~)ԺG{Yn;{Pp7|CcPT$w(òߦ j{ϱv9cfg& & qv#}j7RVzUFdSv5C\Y4#0ۺ۽gw]0/2HEBj]}KNHRF<LPFi,S1+w\)&M*QBr!J"HbJE$dtsؘ_zgn٪ nRMu2P(Xx[D"Z2-#IsjsdB>y ̒J%c2PJQkZR *,݊CUJky~qqs J͖Z3&.ĺfXE6SJ`R:L%-$LKc;s5pfLcnfpfUd*eZICZuՇSº,hTղ喩y{S9ܻE0;hRdu%Ըѫ-?YShxQ]ˡ=[_ _ V@^QJm ETW1ׇ,1FNC V%Ctȶ*-jR\.Vkqs+bDMM[Yb9ҵH%9Rl P$r-PRPμs7m8эa!,H*} &GHHI?b s'4҅QD 2RHj)RBB }RWiCMŁ DԨy1s (EqSj)UZlʕIscT%gaEƭʃю̸թI(`JJEv)v@2S0AwF۠)G}œ0٭HP訂[ 8gԃs(ngU8:F5&N<تJ\#^<)M8ӆo5%om G(ٻ6r4W ?,1bzh`w0 ۍ&Xrn@9,YDvU (W!Bh%FC=񬓑$3:wGj @s Ҹ}V#TBVڰ6AE뱒&dsV Sp_V'VƢ"-Sc2΂ЁuDN6E̍z6ҙcdH>SY!MQYLMm XX#˄b\E/KPB]Y ]T.cHJ31igUp#seCτVtgz:gRidScuk܀ AwŘ{,#„C&LY]Mp1Δ2iQ%@ `*7n+L_ JNL6LE8)ta,t3IYRvH( :V2 TVBuՕHP9{TQAx,Ǥ.L`lW AzmV!ĝQsG=z:hD+kn(0wfcuDՊbs6VNw ԙV_vnn/iΘ鐜4_=(*J Ua81BPuʃ}FlL|@.jǂM-x{X)bc%J.0A=ڦXHGT^HMd '^bTJk:&N#'`A%G:_h] )xHiD(2ӒD 3:Ltn7Xi:ߓQX&/pjB݊X!q 2p4UUBNe~d}N3hyvwf x6L1Je9JtwN֝w۽joƝEzel <`)VZa9ҘGwPiE09(P"QQF݅Ze\@zp/AUP@R@"ƤzP*`idCUxVBh=U;&t :H:U RFM P;ﭢ:5fT @ቅB̓,(*I48sbIنJ3KF#AGWG1A%EF"Z.'ʃQe)ǀ2C? b+'KBFfc-eut"jO F)852 oVJPLI꫷D3x/8TF`!-D͐ ǀ$@}^~MXj h}1RN6V^` к<t\LH:ti,o8W+ B]@4"fgmTLzoPpoOcPHxhk1^GqRV NS#@y]}'9lT.1[pPaPDNHw!!IOG>+ wz\f<,@z|,PECz{*T}fZ :*Bj5_y#h&Sf޹TD9$\; 9mRDJ!Q [C}Cy-"j B,Z(jw3[ͨsP` Q"UiCI9:19QC JTFzP.3J\ߙdü.fT 6h t-7lwd`Aْ )T__B&Kb@ 6( BP7KjnQH(-4E(B{)AHvq䈚H(=GPBq“t!Wh ueңY|\K1Зm 1PPq*~7uc'.EHQΫ\|e}"CPd#T3 ՠYUB2+۵a>_d< ¡D#C<|(nn}FqzYYI$2XJ {jLM7{|2[Iy=ic~:+?v>wjҫUtLR-'ϯ6vwO6_,q쵃h>p~ES2^ܽu] 8qC]x"0hkQ%Gc\u ҍ:ͨ#u=ͨӌ:ͨӌ:ͨӌ:ͨӌ:ͨӌ:ͨӌ:ͨӌ:ͨӌ:ͨӌ:ͨӌ:ͨӌ:ͨӌ:ͨӌ:ͨӌ:ͨӌ:ͨӌ:ͨӌ:ͨӌ:ͨӌ:ͨӌ:ͨӌ:ͨӌ:ͨӌ:eԑP:lu$=r:QG4@rӌ:ͨӌ:ͨӌ:ͨӌ:ͨӌ:ͨӌ:ͨӌ:ͨӌ:ͨӌ:ͨӌ:ͨӌ:ͨӌ:ͨӌ:ͨӌ:ͨӌ:ͨӌ:ͨӌ:ͨӌ:ͨӌ:ͨӌ:ͨӌ:ͨӌ:ͨӌ:ͨb:;+ɨ^wx:HQG 7h\u~Dtͨӌ:ͨӌ:ͨӌ:ͨӌ:ͨӌ:ͨӌ:ͨӌ:ͨӌ:ͨӌ:ͨӌ:ͨӌ:ͨӌ:ͨӌ:ͨӌ:ͨӌ:ͨӌ:ͨӌ:ͨӌ:ͨӌ:ͨӌ:ͨӌ:ͨӌ:ͨӌ:/ƨw WY] Zح~koֻ'S9-VzT酀eVh*,XaQKdeKL]ꪚZg)?:DOuϮ e푀> iK]^(W?6/N.u/]=x&zJzEЕ|]FWO}Q.U'? 0X h%w/*J]tū]U ]{xiKVNW~Drk[RWoݬv?=UwOYyZ-|w~v3Vc~A޹! ! n=5-rzkS`\MD?EPszvg0N@l1 Eލ|*^-wn ^ .V^7a)#d#NM^Jkԭ `r=:?2 [.bav2vdi".Ou+!ȵcWZ ׸cY+/%mV^p4?|Urw6o/LvN3}pNM! A!x!f7}Dŗ_OOl:e[7әVe'14jR%#z%eDvƟG藟_O'Bn9XGm$,%)xOd X QY[Jq7!s|`ȭ y`g5s.XВHENiS&j7w_ϷaavT/̋~??Eθ gMvFr"$XYrVl]>0$t !;$]ZB.}πK#ZGdX)VĨy6DvDO[}0$?j3.s&@qD֕6]:}~ l! M,CYkNJTK<2K=ĕ$;~c> yˇqC^,+9>*+gˣ!1E$OXk)4%aLWŭ|`p5=|ߟmr0m 2~ a!(%WQ7!s|htED8J8/MI3 "bFZN&_x>݊o.}*4Dzj&iGOsul|s+.[W?K!h[5dq+{ gU${ݤ{MnigܶVd~dKrjCVRR55 @Ό舴1:1 z<Ǟ[gr\;xD"GI꼗aJQh/ A,U%lK$r SHϟT (ᩦD' jaYMnxjr;ޠO=O~ :.;$*cW[*F8ӹsXs3w&=Ш#4| ˙>ful4Yͪ+}q)R;R:In3‘$-r\r%Pڤ[:$G!m+`2N%V#}P83PħDl`U NT’ wL/$Ew-\f!c-rxD+:;h$/ʂIub"fY"Bf =] ضd1!VOEUmdbZ5_bw6A#ֆWFr6AYN >>m,:Իɂ'~z.Ho1cQZwpu7| ÿbyʲ,ו7̊([`w-󽏕G5ޯY R1N%3o oi^qd9G0,$^:͓\(I9r9v%[I`mU1!F}Ѿd?lGcy5E[ V(cd3sq-s* nP&:t)E QJ̆`0wq"h;g*dЉdfWn-anV@Iiuإ=йL(ٙ Lzl9eJ: юP>fnX;pٻ ϠoA);W-jS/ܺ;RaO- J'#MLΙw5f {mLFiP6h/Ї\P34bAIԫZksRP}{rٜ@W.fxBK쳷9G蒊q}IJC|]4T2S+Yꔷή] Lx`G$^7$*ӛYCq4ɄwsSvJ@*qoS[p#3~¤O46.OBYxǗrg he3tTez9S^2mټ{G )$[ƌv1IE41,aԻօRjgW4a#`\۝c-ǒFE|*b^'BDr{<'Bq`c_pl捩s8N@r2w5W:Br6bH+Qy|#jJ׷Wq;3Tii߸LZ)223a2{WSi@F#? l#*@R:f4+1#*rVU@.Uf &s)IOe@ 1N1J_&؎Y1W>j>H $YNGxXʑRƉW}t,<-w;MU~Vnpbsy0@EJ36kl tht}^S޿ Ϗ (ui%8\8oo߮gp>C%|2z'[~QSЫe i$њ+-Qu5Y`軯";ɞ3Fs7? =jS9MZɷK*`VBC%ZMcVǮ%zK ,RШJq)>^SQB(֢wc/IkӴ.gv?2l] %6 eX$ijNQOL%g̢/ClQ _8/(Y-a YZ `p"JX`VBC%ZMc֓C=#w kdv;ElhFg&jSP-jўD`F2szvM@Aw&mM1}`RЫl4&8$VpǽN1B_%Sw!#׹ :-iL 9 vfAtܵ@v` I.#IsnpXi),3лCEN.@$vJ~KΈ7:GOJc@tBѓrSo Xqo5{ h5I$J,X`dx8༢kp(c  8*jH]s^%z92^|Ol>DJnD~U?"m$62/p<߼>~شz! x^ O*ӘFLG^u]\ Ժ2ѣΗBm v5_Uů,-J{+蕦ôi ;޹x{SxFe }"^-E08"L 8 '5j)LY__BJDىR('>5yE`11cǷ7K>zzg J4Yl2}/<n `Qh{!'C_5}Ԟ~wIok;nOa س\_7A<)Fc`zJ-w/3!1Ƶ eG%R"oYɤdވ4JͰXojFYvr_6]=,pWrJjsE-_xj tcŊrB`Bfu5E eųjcVQ]-gv5XyUl\oU7ɻE[>bʵ£ b^nZ&_Tn+ەX;,;O历k^߾cgޛr%܅vBN4;b* |SRjNA./Y6wAø 6kzT$2 | #*ܯK[Z1ܠ1T`fRe9&;IU]2݈>Φ)'ySNȰWaj0abCthLzJmh4y)w.Nб hz afz(2_p"vsuO}YĭQ"w<(g,cQy^@icvI3@k>^@fS)f9m=+e[R)Syԗ(Am|zL;}G1!#96%nqq moՍmW-G"P}3u8v8@6s䗀(.%-rW(9>|>|>͒+]luglsyiwnx1в nVt٧c*(9l>l c?m]hki'm_`;~Ul>61_@ׇpߦ?!|nUjϤ`z}3:Kcg&4m#$u3CΐF\")K)3Y ait]K$=ު| HЈxM%J!kݟs!c#޳KNYțg MZ `苘jS$x{E zMdL)!3{T$"4Rm VW*6(6g@06>!3Mx@̮ޢy+9C2IHNC>?97inJ0RrgndģhJeB)B:d@;Ձ_E-WQ-y(l r{ uZMzyaSpJuxJ8H">< ׂ^Ra,f}|Ŏy6E6)"4 ;jJJJe^Y4Bb.)z>`ln!34>8qy*CY%MAI& 2V6:˨'$Vp41,a9l]PbR,1Trl J͐i"ˆ}X}53b+z~ʤ~1+7P7B06wFeg {{e`)bޏ{ ,^S,x ԫX (?F~kXkJÔ01:ъUgܧşLͩɝIDWD+ګ`3k-/>se:8\A Pr5ʳ|ckSjͭ+Kg}k+W+'WNf_D,ڣV0 d $@AKYM/l]?/lͼ|,e&e#ʖz/gK?q:$ ,/vUfv/4TH~zϟ~)l~62+0$MSGg8CFy#!ʭp07eoVvۅ-%Jd?8}~1yp7gtrDqO#/;B['wFqc%jQ)]%k w.[fʼ\J\tuSuJш*JU`wה!1"f%M-QS:hYb9 | UTT'R#BFpTe-˻.[Wg;TʟT`|vlk.Xp`UKwyan4"VW"baJ Wu4cu4AC+F;^'iB>| ]L"Hg_&1kƫ}mlc#GapcwYneEݴ5(H1AUSM=hM`#5Ȫfu+4)' Gbb"BgbjiCNf* T@-Jjݟ' )}-Jp<8cofyGᦅ/EyT*u@(M~[NE$< @Goo+{rOqifeTԍ $u@.p蔃SM3Bġ>C\Z?P˻RgnMr36osq86Ĺv]TӣkEuZޕpx~ͿȸF 8Fl/7 M-|٦ bSxGl<@#bV$ɿ܈EbH bv v(dʸҔTsz+Le?ҀcĚ S|tsQMsGG O9Q=` 6 7N%.[)iQ߅pXA_ #FzOG[[-Cr%fBAcka|vh/+QUY4'4IߎJx8PF*Óe깸IU-G[9M$I,<5箨D ӆ;c`H[ޕLUv]{xj jJ56\w2||Ivt1P,ϱ)YSMNTs|Q/Z9bz4 #;rAt 8lҁ!=z;JӍtoocJNk{`Ggŭ7,F,B h*@3rEmPKH0e8Hc{@\Tlw$*c2Zkc"k WEՀciUB߮m`T#nŀٯ]^;#.-¥u c vv.Mrk]Kp+27oE^ಈ !P[@Z/m8fJ jyG^L^̠*IpK1D^P}p#vP)pܦDHi3 fӭ8s;lqT"F]UD-HrգRߍQEtݭ [(ǘmO-Fc7\]w#z\|Qb77%n)z@E#"1CWkNSiخ&jLjAE&lߵP Ha RK0C@bRqF/[QWP˻\E-ouK6IAA@;bKP(U[;qPG-ZPa 8F&gWТTЀc9 "覶cT[ct:||8FbKt-F41S8\2U 6 F S:溽v]Е2~V񏋬X&-6ϪfEf7i'YY*6"N l3HUˆV1P&'1+/֦rˉD/}iZpI'0?m_阊89 c`I4/r^A8x2g{wOr53h\5yd|ާ&EC`򄖩 l#hŠX$_~]֝X9R»Fq6 _ڤ%T\%͗}Q<4 W*&$V*$!H$a8[nGrf@Hكi=/jX,՞^\窉_9}tu6H佼p&spN$Z f|WoR|)bV._̼s `O2f ݘp.zECu >ghdSqfIcMRi0KHpEqW/Ve$vu@a!6f,MފQgd{7^ktň, +j ڐ*ɒL=\ WW+6|ZB@m ^r>nOc R'GyVUY]xs/>wؕD1;fB̋I"5iW6aeZo6Rm%g.#P s:Cz `˻3{eO>peɥtrPtp;_$1<+Dh.Dӊqūq__Tk>|{XrWƒ_O C)٬Ŵ\ӀDQfoSBa&1Ip*۰W[7y j쉭ղ\gZOc'.Q:3fVr U d b X$A4ޱh,LG­e dtU'D? 񜋿Nsc _,^xȅ1MB<A%$m\bZ<շ; gߦ=W M7pdy_.-С0Fu " ([{E0pۖ^k r;H6:CPX0 n*W5{ĐKzI~LAǮj~'!9eʦp`sV/ض f&r?+%ɡtX/:L>Ph/C#X-*ѳP (c3W"V/ej`u]v|C}/ g«5R`g6R[5>@x͠Nl;~R|wc_%u̍^[1.0XKf(h,'e<)ff4/z(9F.6Sb̯y 8'x=rҢq7Ww+.H$\{4bh/Nk Ϗ#9'¿xw7-5 8.G!zȏ*6qҘrF,!5-mD;S`c}kl&0 &b!cрvG?M_zn{?0cػ(hI`Vj ' 8NWc*qsHZ08_X݀ Dȟԉ,_әd2Z;@0߲ I4eO^4̟<_٢rYaRrbrugVWi|~^mm}ֺx/' oT9/c'rԛTkmN%OILm~]I֪^_=I]^IQ~j{`oP zîMԀ㝤= A߲jmt}GvX.Da_R/b!ဒo'CZLD5nqqa{7?[\ {rPJTG}97/'n[H$!To ԮAcK|wԐ<Ft Eeۛo˽6(.EVCxt >2yP(Ajqql7ɒu6-^[W8}ɪlɷ4nj` 8Gh@sZȭ ȋS2u1!ud )'>d^ZS’p]{SipVw 6O7-?P]a??7 G9؉YY,[x\ O 3/~g@9^,sjfo6-)6ep_Ϳm|43ydy˯PA^?!ǟ,wՏ4I&Md}AfUtMsx+ʯlv޸IE)k=eD&EMo\6)`? M>>?gYL"RygY_lqOڽs#2So:;m.߼֠?Eu}ܛi}MmJ'79Kˬ(u fl6{cP_ن0jK^;sg4/Sk6/ O* Ta5ZxGO &sgHvW~lf>g[7)I^>ؿħOpw^*e*cOZ gu.>A#&E/ߔ ʠ0uR݇R|=7[F}z`/߲O䆿CYÏuN)|g7uF)b "?+v?}Qϻa?uKxfDy,p' JF"*J3=%6`bWOU׻vozhWww[oclm YS)t ؉ 1_;"큿0Es/J]/~oZO`-/֦AwvueI7VmoE/:5u=+>.Ӹa3.GW{݃ǀS㫞=km?~O SG4WG ۷% Ꝍ3Ųn|rI:9vܽz q c&ށ;xtx ?C:t/53LĿ}'.RPI@<4!Dkt^'?[\xҀ| Qz[nb0>1uOjYoDL]Z#,r- ỳ{W- OOCoYg) xq{Q*Y-)o$5D$AfEUI˖[aV <ij-9tŤ'ӵ: OocSWJ8.l"@4Vf¢Jk%LG|69A)Ab1 Gv`4(N"3Ƹ)R̊{ߠ"uNLl9B)%6p#!6At O(!>PcUcjVOh0LCd8"Ė.UNb๸g+bVr(!>F?CfzQ^Rg,wgcҚ~p}w C<ۨn)!jrgcMw&(f8.xxPQ'Ѐ SPSL2Rol`z|vлk|k`mcSPYN%Zkb %3h2;@ĺ<Q"jFLۍ:.G8wRFr"Y ig,ĴrڗW"_W@Ojy>0ό!+b/w5v;xrUA ʑQ_vr_u*LhV_l]=wzs;R3F1>S"y\7xq2GF`Pl~^To=#QTcuIA "i FMH7 C2RX'[蝶D;hK 0:ncQZVSr/~pr=rT_-n_ HK< ߯3oC wmQb1~ĠPm1NˊiB{IVMWtlH6aFh?#9|nc̦$>ӳ2&z]s&L-rA-H׻3ާNe+Rl͹#&<J= S+m-Z1dR53$Q>X-tJm!eo,art @)3ڃi$N'JP.q]Q/oƬ~=\Y[/E׌4Lžl=>sN[{l}s3uPO= y1{o){ńZa޼6xF|zG[ӛԝNOv{ɅnV$Ǻ`&oQfIdc=%|! >>2c&zꌸyVgˍL`u Ax`Q`nƎG7Uvdu.5ot˙ >y`5)Ds1e8 8ǫpLK*8xd5s.ҲZ}Zp8٨v.Rk#ƂTIaJ4pgh8&&82npqUu{Q< G11q(j5.T)eus!J&5oypsd(QJ.8t;ǟ Tډ'&3TQqJ1j쁛t,/51FHfw~Z \;T!ezukQ/jVKBzd%%DDz2%{JV!ƿ$#G=0=cbg|əA\;9o7T?9R4&j2&AA`Lu6 )٭3y\ 8KĐ%M pBxɓ޺G^΀u?Nw|kyoLffS| _4EsRV5 )d9D2cb9iAMī}Z@܋t1]|zxI2k5x͊S4vKyrKKy:+>[}Aѳp_ [S(:K-pu-:0;_{S/S#$/ǴI^0 R"T\J;ɜV1 fʾ[ړ|4ǹQ_ޭ77q5Y+K$w_=|ޤ>Ce<¨n8̜K7=EJt'Gu;,;PxG{X*I˘e:(ߍ΁WOmr욌b1̉4g 抶dǥW4v#1K3ReTvsIoO>?Iy >| _;>!$G>C^q^{Odsx~ɞ[x< #ē?-y2u wJA-k(w< ©sIKiibpNiCXD=>OI=_`RoG'gw3}XU%8]{ȍ/Ip r;A,y=3GKnZ8fU~EkԽ3y!Z,^OAc]/)fvp#[?^=[u7ݿ(.Q-T'!J Fe5`݂kk`$_?{]b:m^V7Xˍt*fPW̨d2Z\=M4UWeczeKhǪAuP&xV4/a @R[;C]Drp-U]LضR!D#s¬Ij q@HݐfIJ %Z)(WD3DŽ6skBL{=냽QZ63pcͪkS? 5n z".آN[sO=w$<xlqnxduB,HzG k}z9.Sڥ]vp .VDxlAN-Ȩp- hhdfK*60?fRMȉ7vdJFv>e?^oocO0JE s' (HTNgX/,q93ʾ<ںßfg+͗_29ϾM39]oPN_i#l{a//RrRkƍ/4\NU3_~{_da/ooտ5yݍyh?׺oOMYVw[$oό& xKޘ]<[rƧďcC'.ZIDL)DS$S:RBF ιvH)E,e v~|0?N , ޮ: cSY}g? b֍Wɤ쨞)rb`%eGYlFwv:<]!r12~vBhk?Jrd}LjK/ָ\mA[﬐nQ(7#bsҟיe/{,G\܋Z4{hu#׭{KnwE\x^6fI1|Y͟^'"“_ϗҝ-œםy^gw;*S݌WMq?1]|ߴGCromTZN&!RIf]# KV&nF?y&qqs} {`_~d2f28$ ֮F{|FiiQ &&O.Խkc0Nr~8!BtdH^=h:nUAZ,^) W6+K<`'2GnኧXD{ЪS8XTUYD4\*;pI얠IQ5=dHdeA`3\,E ?guy)˖5#zqr* nTsS0/F)lҔ` q : 2V:l""I[MKh["`WSxmKYj,Yyiw:c#RxĜ6A8D`%K(%7]qJr|T!j(;w *6V=)yeZc 0GC}!x<4)Yj 8!CsbhSLfy}#Ĉ>VG1ĭNpBOIIco/JnuǛq<ֶ=9lтm8 G#=^~aa Ig~Is;׮1}.K1J#_UڟΝp'Z)"b-NK@T@)a$y:!>c%K=G/*ROd~ l.sc4p F,fOSu+w{xX@ !K!I Wi7yr 'S={x /V[A\^Jq;[~ܜiI@4M&v> +xt_s2MgΌʨ"|Tê?˛g_'n|w%>jڮd5]qL8z_g1Pkxh$#|t~F0kk٠7ޜ@FN1YhBJ‘ cQJT"% 04FN(5AA溹54NÒXh\'&.< 9(栘S-z>b JѳR6ؑM"ұ֫ LAȤv =a #kt$m".ff*wN5mK'Mq$9:M!4`gط zi>4'kԇ[dŁ67iup*!8@t"J(R ,|7?+X5+]&mJjL)OR !R@0XyD$0Xΰ3fWzWt{vC;绀Z^N:kDb4Lj(+_NKp=W"eiA! ڽ-I = !c7e,dg!zlhMGOlr /{y+[s"E㏗sr:ά֦!pš(2Z*% !"#CX RPYսߍSvJD&P"RS/NhKǡ +E!L1Xurj.Z4#4?lzPikSP-W>ywt)gTbS΍-'2̚XBHt^ ;ѵbzA:L ;g-'z|_uBgE@N&x k@` l=|%X)QXptJTJ C0b…CDX"RK4][ V i)Zq ޘB⡬Ɩye26)Z &Z&e2ؠcoHI2CRaH(͉Mm \kgaX2Ht8BZ6-M4mдcjEk@B"IiD.bG{p|rC?xn2rƳzvxvx;CgjXjgi @"9&46"k+fZBs@HBܻL)S.JgV'0{͛ )ˆa)j"Gѝ,R,*B=81&[l>֋Vtnv~H;8$jee»V"lln'? r˙~p}_$~M.~Y:7cz<$9yy)c/uEI~h1;h+QgW|di*Mpj9)w%_x߬Pە&Vn»v|SLr=-U?[uRubT.FX|}n?GE7哲vGAŨ騺 g|(N(2r~y-;:~782[롆_~ٜJ;߿ ֬d6~PA^/MN*/PK8tiij,w1ij?K;w7,yP\UHᑸѼu9^M5qim$`A3ϭ5޼fxl.4h7qKcJ0GsRY 1N@f@pMD rTfFL*>eǜ&nlEQ)0VD񡥏KfNŐ@hQhF2&IlkCQOkfeH7&H$w4^: 'C%' e\,E>ɯvI^~z3bnj}ݼ9kA~G 嫸NĀjo7~) iڱ30Sx30͜Eģ5|nii:ܝf q_ޫwwmWvf~1`v%1Ƣn"e%8)DlZ1 tUu+P0PH/=#Ŝmk +=MIqT4u=1#طI75{y"{RFkC-)?}u܏HiX#~C@q-`wEÁuk*<0pͪE bb]Y*n碘2k]y7>~ȯGjL{[58\Q#o$#xBR)rD[CPm$Qg暷hɛ9Av=/?nt5hЖVZ7hgM9B .%֗p3^ZT~`5yfѯhћzI\x].̾FW$~"ǜJ1+ X(P7=rZD@R0|[\T0AG $7AI؇a^}i?Nњ^\!H<'{. JE?_S*! >xR /N&3 %Ƥv`Fq#ߠT^>qm'J KL3cɑvFs'A{aOlji"k^V S|,MX;쀬:T<"ۖTSJŜ`{cۤ8jhL;؜Zʤ /nQ?cZ6-i٠} j ,vYGnI]8kS;A|8 ݬ 8fE=b{ö[e P5bgMy1t ۆS~yyKZ t 8FzIa8SZrc^5o`qn"}פfè0:o o0ylx"S$1O5ojME=v*ZMGR `7A9a# wڕrd ccPskVģGe|z>tjax*,Y%Qߺo:8Zd搣`"Ysq(`1/&Utb5,U`LcC`ܷFJoZ[,j?,Yr0R}c(成}A0u2FQ4M8C΃b$ ti>!j/T./>nV܀ #)xK1l=(14np<*OjqdF3K }LAFvb8‡`:bL]Sd ihTv!^ԗ?uچ|-in'?]};Wױ]\^xk}s5vIWZ>w 0۫y8Y D$6kK\,?kxk>bRzަ "?65GgZC>(}wXccVT"a0J7ax 2D_W÷WQakO_Ʒ߽Sj[h}=o1wJ5X5B(|k ka)h&>i2w(H || -f@nðWQPAZP]J"R5\H40/kϏw=r\<*#^J-+UjqӂPd% *a5?O?K&GepK°[a{k!\sn6pI] *b7ZLr %:x(#?r܈vYWFklB$JvIGp4L4'nlB 'i-)Cv +_n(Vx>AQ0셍۝<*SI~;AՎ {崔'qHry~|b4I,CelsbyfyNz5>'] *.,=t.7 tMaO"~8\Ăs笹Ie~,씚%UY)Z)ͫLKyDLZ {Ȥ ̄H6x:ZJ'fA.jnz8JMBP)Z1b8r28TJ25c+\'[m,)W "z`e>"n"Hd"H&6e t<1YN8]h̘Tun]cCr/hV~[eGew 'Y:^_8GcQ1?lIPW[ T+4Ԑӫvh8EAI|9A\[ƅS~^/GYI #g >Iagu *uQՑGepL3F=鳪N:xTazy֒UІ^tBVV` Ygg%%﬿qfM #k^`3 3(hg!)`ɍsFŮ[_YVB(Z0,KiA*(4-S^R NB݆2Ұ;x`zBfD{c!y'C`v TmJiHɷGآ]C<*3 wfN b|%\O|:]o·Ӈ,*%(q =v7UۺɂGep*4;Nea{ CH_1"k5"p0J7oWwō3ab1W5+Jw~|%+Yoӿ}s??GVhsNLca=o'i 19rW G0Gk'QkaOs.?nam\H~wN'c)?,/0G,sȦ?4.0xUnAS%ٯzw?$᛫V|Lߞ- s6'owZ_njX\_RV4~_$x>z1韾GKFhckA7q\9C868š@la'r>wY]H~ ő]5QM`[|7^xgI;_8k7OXG etE0~ w:N>M_$pEɗ`|oy;cI `v[æ4'{2L%].e?GId7DFOcOM]MN> n?;Kw͇/S^s|{-vXżϘfYdg^> l]m44ŷ~2<Ŀ_N z^Ej>+>HɾM6'-EY~tLk>Ʃ~-O >WC*b57 ` &=/ٞy+ ,Y f +mHʃ&-}0, &̇,ѧX oGk53BlGuUz}\R 90Ηy{y )C qJG]̾̑dE LeIf8E0]OU~y_ eofofB{3Gpīg&1x dRӍD_|BΓMrωaOø\x_61n}22׽t|v t 0,<;ZgP") k9(҂{7Fr9hKEs2?82S&MU `]Cr:v;[7?NHi&xۙY-7o|nx~h#"琾 ?nw^^'p]pQjG"q"LQ&5.pEp4$K5)~eg#z{nPɼʹ ү>_>P>:KVw~E˷=%[PV?}Fr7;xU$#+]0ܣiZH;xNEu]:9>{CgSFV9# Y0J]4m.Xh[Yv] ڧ} Q`2"^QJymB׉K]xWؼE=1:3y".DjAƩL R"K j doU3(I֞m_ܺqΫJoY9Q<1C*&HP 0|Fl:JMJ@rb=a;l>;4{Wv,\mw!/˷Uf "^Z3*ZVZ{^Ϡ/Bk^򪬯<WO1!H)zfl 8 TR$Nd 2{ZDS?URa ; *lřk#BWDX.\b>屷g̷_rV1ĉ2.}<~s l4Oy0~< *'0f5 _8i|O9.|>,q<2au|C8#8fytrAvXr&q>*ҸYӢ5vVț[ur>G',@1x!㓿ER~ ޶l:9k>ɫifFk ,2\oiM3'~xr6'j:OH0@y{f J2X'A4 ?W=h6V^7mC_o)L<+|4܌5˞m}}߯ y$;; BWsYɥܶ&.qӦV) tz o+rf|1 "q#h1jG`_ш '|6"ba {H$1{@ J( ( Hp,O`-DEC ,?tf^5ôαɿ1-.vƝ]ǵ^I+~@W;+IqKWx7KHGN9Z؍utǎw^/CEo9#R@I! `њhEQCC?68>ذgE\̔42VCX#'ZDw9Ld6j4棚cVbm*I| !L)%Dc=h=p{b8<|sMSɕx<6(GbQ_$ɶg"vPJBAG@i)i N8Xn혣sc1מ%@sBHH/x@G.8P*$fEU,a@Þ% |uiȣ^QޫÏ^Qwb:ep%H&L XHLrAG½J#Xc@ػЁYΕ)AjÒru|ɽUeS35&}7U)ca?argp:5{Ϟ"0c)>_/1A ;\3Q#h(wZy4mQ!Brs4\_Jܫpa]_ۃ>Hw`gk~|CEa+rr}\+0bAڰFBN sSl rЎ10.oP;=-1~ѠːK;l~NG3Λtl~d8a_Iwl-. Uܮ&81P̻דxm'ɧZ.?zbz7S:>2W5דזt v>O~OyGt l?b3i8WS8ml&lndv2%X6Fṟ?#LA;ҐNzI 3Fb8B^X=Yar{nܢj=hHXm_J~XOvmNR:/dWTJw_Vf?f2 xw-&L.>Ihn]9;I~ ~ӌSǃZs!d sCMw搚 w!|QyW-kH J" `t/bԛ|[R.Fmxd/ɋ( EiLki;xy4Ua{9u`piuE-U-!Q' Vq"H=|F|Y2sWǨW}HM [BP-p-WBݣG=]>}S#/zÆ?0~EOȇYt?.;G/PCxJK)OEZ*(1'xY*O1ii9(IK!\aLY (,Nb%jO:PxTtL!!l{5@NI^S"HG 5Ik $!2Dk1p4L| ƒye)v8pHAGVp"c/PCx\;QW<␱6%>*HV (T~[gSRg8f(q YI$I.i-PEx- c HH.BY`|4/PCx/kJREPr9 `6Y֓J(".gLXCA|֛ O~ 5gz-"ټ 3!',Y }S!KBƴAI{f"IX̦2M-&v>-1x)uۘb]15h.(UJw>-2xzԜ9xD#HRzt]v ~<~Y%1XL˲#O J"<熳g`aQ#q8ĺko>->xGф"O}p{-^j߅"`(Bsxh>_//G?N9 Y]di}6 ^ ;e.s1F\`р sSɾ_O3R4G-kB+*z]67DEom@<_MiE:))`QÝQCvj)-^1&bEQPqO9{  C}Flut",(M>6gS;duDJFW[3P>W )eUA~"s#[z]_TZO7iSIF]z\Amx΋wxOpA$0-![ aaڋ^㢽v~诲<IxR.*$%)եȋ54<Cc ?SF9LICq\]]x0M~y/4ŵEDO+W>`Hʐ$!bг3rb|Go[IyP.p6F| LgE(/pe3oYr?0VGݼ}Ãa~JQ2 4ׇWw[wC(7?:e ,o>OW[yG8Anհ 0O"vGlkp3WL:Al\-76VVū:_gF~ h¤\O.ӈ]{PŸ&AjsR)KʇyhAjIDhWs!QZcr93?u~g N]>zثQ㴉wίMf|56dܾ>o2¶o,2l : ٫|H4T'qnsQu74CEἼd|sL: BYSoSSj<_ަ E͵?i0?~X>#|{!{!=uwGf6;nzl >|_8GH6,޻LHqɧR hMˏf ῦ~Fw?~Rwj֏VM> t49=9Ҿ8-6.|B MpGGk?>iOW/AȨOGfsњ۬clZ߄q$, Thp,T J)T,)EC;|<&jBp눐Q(2+,Ƞʠ8sr2NQ,-)dܻjdL,Lu6:ϕ;pV/>zf 7 3N:UUO95ΆeAN K0 (_RKSV /mDI#ekrhnHVJ5=Xt>8*ov6fW$F(C-jՔѾA~?AdUv=׸FR5qWb-8C5(-~+^}H*(g*L+LtVRt,?idK&+$FK`u4>6+_PɪYb 5n@jH*mBrEv=ҘbS(!>Wg*uQj%ŻŅU.=죊n5,=45!RWt8% [xye UH*݀Q!ay^4UHodY#IjYp2Etp{1y7$k6%orMnk!ָ\X [*"SP@*_+itt85Ÿ-ȹga|Yכ-㇞y/un Efg?ς)ΐ_<4; 眠Y oT>dfB J TQEWأ@^8A'ΰRQQ,@ڡjp"g]zgIDKy>J܇D+V+- ;,B!7:]p! ʔ}^ *r%:D::/lc$D}v8op41^{xocd[Ljl  s`&åτKx~|bn@^LswVJTXĢe*ti"p(2^KKd 缴ٗZ"_7w1*0AJi!pCd: ɽr-j:*} .Ǜ5ɑ 砧Ok!nYXvy [흺SPOYtf[ˋI!\sfyie/3\6^`w~4W{XK^[gsQ.׺ ifLhpÏ/Q`VtY'7ܼHy 7w. -=.'臥ݜ5FoF#/8\n$sdv ;XE֝ngN(P:w zavȔԣLjns? >HD'[m8Clr>i^^N=(u$n?4$RIϓi󿛻4]!E:d=ղF;[džm}F`Ѹ=6swo!?{w4w"Lfl&Md/Z &;J+ƚ:цwy__}zՋt+O'-NZ|?]^N-wʡeCX˱fwϔ9.w_N.Ʃ0 HѠQ S#HgA/!t[b_Iݓ/a]*oٓպ쪞zg]q5l"Ȋ[n1KL. jSOR=戸C]}\J-T"CK1cpy̨s: {e^\ċݮm>|q%DE{nCQ|>fؖ#n=M#xwP_zτ4sz(AY<%6+Μ5,4gm\] 8+*=:zV13"ΐbjJR:堭12@32]+zUe Ϊ;_s+o/J&}]fuM)Q9V}=,|=g2戎Gs|uWc9z _E4; _So<թ?,sAdF #$j)l|Ѕ)I[!Ѯ6W-gXʘy1 Gܲr/"Wx<t B%D"^,v5]GXJ }n<_–J1J]t—#s 1zPZ|d+zP!GAĭA,!RXDފq DkD]JRTΰ$ ŠSkyQ"uM!mUK'Z]T8jhv Z,JN$yMͲQglNNSUP۝1+hq8Hxeu`4wwqJh)2ĽTH@eYHnI-(wU#^%ܷ{=/u}*gf83;=?3C'v-{mC+y}Փ Sy+j 5֨m]L~#;GQȝcmK?y77lXw(OxzS;, 9qA9mVϥI0 *>G_ŏGMsՎDF&("8J#(ATφze.ECfmNqkЯnH+-BWI{F Q*=eR-SeL2[zTox[t,{;V_#C3IL!@J :" ^zbgaNo"Ã;/9;PX=eL2Szʜ!&xn&`|on.dT^Iz%Fe Fc:˄n5Y"Y2x ڰ#+LṚygy34 SͩUC>ŹMb.\JsUs/5Uab89@JNZN >NGf"Ҹv&ARx(Jos&*̺2eւ8&2'u):oP`ErTj4#+N-Tm32k& 뙱1$Y&q"Os)&꜂TIYuf֠ъ~;L MP6Mh1,. >*#>#B'y'o#uGt=È7 ȿ P QN#NG`":rfף @+=UpBc$k r!`%G .%,5(=X9#wDs_M1yKwL_jgΥ"wl-mBe s],"RHGFf8&>:YQ!&'&A=p$PlL)Їb$RG#MKb- 5rD3T9KLBvo; h^yr֢[vYoH(&_NJ忢f#BrXs4_JO%w%) dkM`%m=@yQzrv-ah"t\*"琺`y ]q|9j}XmEj>r{{8h 2Bw7ezwq؛Bl!@|d^o0ȍϮ[onpQoܳe ]pI ?8)77ȕY(vy./Amjw'|Zdc:\2]ޞ;EYeUnܜ,)[`۫ |SwA<˷arz0OkD Z쮪ݣ\dC\rK󌿫dn*U0'k̺Jh\}:`Uxex2h-%*av)թ'Ұvwp^EqN%t\YǏҞU˝[yO[g[8~o+  zN|f>wD]@UtۙmҬHnUx؋AΞR&2q`,<ēNYds'JQvĊ:QԤ8Nx m}ryx?33{=WQMA(&0pQkoYg5H*pUٚH}67/Lqܩ68Xƛbw=rn;⚷q?# l:y]uy\\piaV&ƫv7W>&Y$i,5c"I0UhD/I,WS7+_p$AZj*0| }X VBT..M8VTVNHÂlna?4zm,5_Kn^]d**ƌpW, e׊,P%a='KUx 쾀ަq`,M^o{]+9VPJ,|[zofԿ(ʁOq3osggz"AN \Måǫe2'Pq|Yw|ǿ v^ҥq rLb=sу.l1+YP[z3>M"X8m6I*nnjy1,½l` 0v|f>Op]73ӛp*?ךo{жK~O{{񢫙H^f8O~k!>v d4XK%Z*EföLE B6v LU_ @;)i.*|҈""ƅgrcfz=dz9ǥZ|#rPO^~t}~}?̶hN7cWaU3[|QD5N18a.x!GBD5rTKuڥדwʁ{cBcj "&B,3cpxMɣ E[#s#"E -HWz#w 72RX‡ź[DŸ)0M8U;''Zj΋`~-okNIqS#7+R YXoa L\HdloHN;k$ "A:᭶&H\ҠYRD)ĴP%ma .֝gh4Mg]Vwt}9jZi͜eݼ2#1Kx` GW0;\a\3 AP`J?"Ir2`d)/-sM~罛0RnJc ZeipX0,g` 5ń=ʑ)l|K}w;ẅ́P @5߱oֿ)/>I:?N1%m6Ռ 6`~zOٻ7$W3%y Îl 40yv&)rK^lE2$[R7YLF^j[߭%q,(Ozlϳ+O_4|,tscB+;:hZog/G+#ӆA[|?gx HdPژ%l=Fj` )f5:9j3FzZ#J/cZe^͟UY}yCgX FNdg! ]'?N LPzPFV'2J@fVDjwXF } C5a|V򝟼kGJiy_/J)#ƀһucY%N?֩]s/7 MOҒo,eE >^EjyoG-+kJ2,s?noXL"5ATHUbS @UcaA!T_a1cm@kƣL3zv/;mtov*O:ws(wý;o3ǑgX>\y.tx]juvԸk(Jh%,kD!Xƕ|-.3 +w;+1+@l^6_R׺-@?\B:{`HZR ]XNcckwA6vI |rs>XUR6;ڇŻ5Jj-)QN nD4B[D1~"X-J<|DףM~Œn}0N,(0#jgq),YZf.=K},'Kl9-DS1Pe2ED)*DmzTYI%6],jUEPihx|Z/kw|oɧ'SZl"zڪz-}O?;|=b9Q$xa.fU13Kfmc8=~t2E,QV] L+Y`#bP,-KYơ[G 9c5QWY\NE]ei=YF8gu%ngu@< rp8?~x?_2LT~u\s-X,}*(U$*^UF#,h ௳5,͙{d`!ki6Vn<ܴL*ťɒHBgG &\ \98BmhpɤZ3WnMxv(őCa&<.?%F*&T` Fցg19Bm0cp}p[ *8~XIth EIRNFjěsͳ>Ez3%O2VĊY0F7F7:6C ՖR)E)JXꍷH:=%ܢ!0"fwwj/k!=b-I^KpXW?~v7$p?RWi"q*)$%LOшЕoXS,axG|eV::Σax".jC$%N7pŠΩ%QP1@C>PP2),8*dZ g8/& ǹk(N0c}q|ZԮl|"qOr @[ϓUI|̈HvPFyF$d)m"w$$yo[=*z7Q]Sj)l,$eۮ2r2{Y098h*gGt :Շ X,=Mo{_-/o\X5].'K:oOaTPQFGυV(Q; D)!&J> 7Yō+n ȼ\e>Ck g{z5mzOTxfKe›[۵JA@=UEz/L먈,2ђ4]"O:<9hW1}G(sGp7FF>tѷʕ'=pr4w#ĭYT=kqg&X1$ h)`*Қ85vr]Xle e J.%}~ofoև3Uoovm=P~ q]Brwq5Kpb&LhF&Jd+Ζ&`xMKB*0 p3Hdi/%& IAE4%v3eǺ=ш'Q&IEyPRH.8!(Q{@VQ9jp nu;;w 62`}[HVQ;YU'$uUQWsm)&NP_]҅_q_F?BJlt;w -&9Ōan<RӥXR*x7TXreEHZWDT.,ph#]8šRS%:7N2*Enh.'P6:k֔z.hCL$]T'  F8BdAR(R^ KzJQ#>X}C利hvq]y{ .i;tt&t$j$<5T0<*\NY%<x*F(CI;;ճK%5I !**-Zl)१6IRR2 kClz٠ fvGs+ܾr0Bnm24KWlBۛέ~ғ(Sx>]up&;]>Ѭ'+|rؾmLJw7>ϙӢƓCvono#>Y󭮮<ٞw4<쮤=]2F3ص4_<q^g^4N6*W3KD87*;Tsm5ݛo @|A#@*Gz%%HzG/0ɬu g)D&n=1 G+!\Zd TV; T:c\k\ "8O` y*Lҋf_A~~l8)m@D%,x#+8BF*gH¿ā16-ڌ%V!ʥZ88a\C)“1MoSm'eo(W , m=ZdBP {aƣ!#\Ck~tUigsqi0|l(3+u.8rIoLM㼱;[;]M_}si} .۫;?;S6&᥊SBͩR@,fH~JDEJMp*8T 'h~+V⫾zޗ0.' ,,2@@qIľɟ -U)%c9Aʜ։9&fM5irsr[YkB>& uOZj>ܔv-xcZabm nC2`6`nyMB VwPi֫)Q2a'pՈt"+Y:tw%n:BrפBMӽm PXWC3X";w^$wcX6u+LG?:m3{&NgNstxv+ْM:<gVKwB5c]Nz+D-ڋfġ?8r ~4Mh9eBR-1c+F *.5xUAJ!7'g6ګv8 olTQ]5GS^$M{ꚫ^W ysuKPđ-p72{> H@@2NƹB >zɂ$ H0;,j?H%/=n 2ļs6QX>iSFEM1@FE+ĝ79M KLӉy9 =1[#VFAS Յ QR\R\ޤ9-1WxfJX3)>1ppQ;鷆 ?WCY eEYٮ-]ƶ! -C0CjPi^ȕ|(- Z_{ Bւ`RWH0=uUUz( 'F*5uaԕ>#JIqqu: \X]\{at*j5UQW]үPWzFet@ j0ꪐ+PUVkWW *mSWP]q|``U!W ]j^*T*'TW+$Xs9uUAWZ)]]*n+`k; u % zzBj^=*T> i`t0ꪐ+PU\=B*%oS+i8}a˫(f>akǜ Yrד%Ogt'O0E\H69щ2IdAS N!'~JܜrZXJ^u&YLYǓƞ?mMBFlE5cyix)cI!#d"d܁eߜ֨sR(D'reP&F*3Eo[%GNA_↷sQ7wTnhP3r';/u^:{K%e(sQK*ŊWz8I(M0(u)DW\yv\eGth!KFi?/i< DʛyBr2A*/c4>y+:jm ڛ:y}OC=_Dc,3F&8笃&j<,{iǂ_PoG&Y+tfK{J|K =wlAdiSn|BRcVӍwH6?rJOl^8CKK#(]?@6a9cR[KC'2pDAibh=9š>([Q ޥh m:(*H=:i#95F CN蹫\!*fѪPRRja?&| {huf"G$;+1=Wa5rNaM~>}\4ڍnHxei il kuۏYagf,Yӗ~kvwnѦ[D$m9PG&㽇_ rxgZ˗Nuw7_?+?\cj;zO"Ǚ~C<ŏ,]{_t.1\=ËsW)zhfƿI{rل%r}q;Puϭ͹1 g$Xԏr9 %P{΅Ӷirp`f-V(q]u3BuaJ# FOH<9qJt;Rm1S֘ʀ2l.*wBzǀ3Te[')K DU)츍̘`3%#IVe 6jqM +BDE"@1aS 6 ۀL2Z+^EI]BP4+. qI NYa^ d9pD0!et}L.[i6"1B\2FΉSL~=~@[`+ﯫ~m> M~Z8?ϻD 0%UR@Ԑ)%as04m=!'י9*,*PPԶt'm^Ut^wU\iUw *Ç1n3So :3H䢤 F6yJ"x sFQe\.`!hN8֋3T[!%BHu b)})&B@TJNcBޣOV\V#8E M64O :8#%@"[?l1/1\7ao3{N] ǘDÝ[I07&ž6v>v|h2CfrBҁz2E/ bo@B@JPw:D qd̙i1`8.(QXH>I[\boρEVB r@L93$6(Y/nY=ٟnFL >x/@9s)P։{. #t^S.%XwksPqmsa.hfZ%wϩ-u`Ni{O!?Y֏QwJObt6g9/${6i0'LzZ  obJ-٤S=8qfīy.TZ3օ褡Oè(jNJ O G󿽱sK!֞PfJJ-$/):&r mР2ZWk- $,2zi}rfזjoWF5UC}4EdQ(`uP>^T *PBJ*j2P+~(]R6B[.Vqn"sp72' H@@2NƹB"Sɂ$ H0;,Z|VK^Jq1~6ܶVd%#y1 >Km̱,|୍hᣈB6*ZeLy[W3ZmZ0Lǟ|'Z/Hw53Qc}Д6B4 M_WmD 8u1t98j'#=paUj((05еt`k6u!iEoHk|yR PoC! ^pG\O9`kW̛ռl%VeI( U`!DO6)2nkwQNT:dM!r~Nܟ{|{-ڍ& >:ʹ2Tp+SN0p-VqȠZ!?gǨ&UCo$g$uOhw';hK Ϳcӛ1{goou,S oꦂ#s##A&%^zK2^'U~X(y!Ē]R5)Z!lTQDgYF e:gDYD2%dl:IX{[_R2dk _w۶N ]&+Z~#../J,0#wy¿UvJ|Dl:k5oЮͰ?Fmz]S -UU0yܾ~.bWm)2{r9,Sr))VJ=0cp[4F3C\ gWsUt{W'/#h奧1#GE2\rzY4&%\bT[\WyP( If#gO19XHQn4R I^kjBjt3]L-&wZ쪛jk[e7ΧDݪEp-̏.ybEl{]^~>NU[=p-8ƍƹu@f͔9I M9:Z1QKTEc@>^ˢIѥhrSTHcL˨jkj׌QEta5x.TMJ>QTmį/?\D7Wm/;{x77?f`d6qPY9xF c!HA)e"XWc #26EN\6AW5QgtM`U)ӠMe]?5Zv5nRi(^"hcnI7A!CYFMujĨ^j2dlSƬA4#$fLD <AI%ŴH&t8h4Pj؞&iMҚ)iHge:[[.9K'䬴d %+ NAOVrRt=_?k9 @eۓ)d91 uYԲNlgzܺ_eLN<&`,x_%Ev߷xtdHn4["YUC~_%dimxT8ʼnpˆ(2QgdPfL,&Ct[V͓#fhG+qr7mӉ%>Y">H4b+3푔6zP`ؙq4TH0ȸTxd 8y-]`~@QAsglR*:d2[p6q[8S {Wk_?7kJBhbe'si9(2KN%"1u5 IܓA - {«aOPk] aV aV aV aV aV aV af)(@Ŀm^ y6/ڼ@hm^_@~@D@hm^ y6//1ӊk^٨m^ y6/ڼ@hm^ V6/ڼ@_X y6/ڼDڼ@+6V͋K+V y6/|XIcM=ڹ9 W9v'1á%* J* 6`yC$Z*o{pMЌa’ jxtRSt*/gH`gISA-qltcJ#68R 9HysQ bg{(^%8e}Tb VS:jJO!6>wJózGus,љ#JяrۧMh TZ1'd`ӨP5:^9],/29:*!!)DK ;ᰉr4(b52<3ӱf/d^~>m36fu峐C$骡IP{שj| b_S9ǟZWt[:imlB Z/umzp=/lׇIٻw]E;[2Ϧ-7+8-ǻ0XlSgMԳlIfǂ(zuVrsptE9I`M:ĥD] uNKIRrVs^!ufuv {goN Ɛ'JDTh;&@b)^lMQD6 b5[YE"R D(^i"\5E t(sTX-"ro0CJf/h\S&NSP'Y0bO)xƚBJ4OUO}RdqbS3\H!ǫ?0I8ZWkZ19sUNk bP%TT:GlYBDIhcN{YU;'1P^NL̔r6ȹMtd.BVA[DX+/hMjM*xjCBY<WDt8Sg ~.b)LS*{9_x;gI[hP L?Ti}^C&ˋ[c 64\a9J%4`8ZS\Lhc hq (AyovOo ?kN͗`Y̵\mO,`1Y!% $!S(D!}bND 5^jyqg"#v&S˻P-iKvԉjg]U 7ߛ}b RxqBŹ`̱j2>OY[EH(򕖜UN)bx]Ҕ(/oKA%0k8NCFݷg//7z8⼅mt hu+-!չB*3hˠ`p㠟*5DUS X~ *5nk-fǵk՘EZO\"ͼb֨J*Hr Yi }wP?[~V6*lUi1HF˾mWE!5Xh-yFAR鄡/?Al_۶X 6uhoMWϦvw"NHkm[pj:R j3J'#jw`AaR}E+9ZƃVIYT;Ml7*DVgWcm שn*{t_Z;vLғLz`ea]鳫`R9sFM{f7Et.:@v@=wlao!./ kk:ЌR0Jctbtfkdyd㷽ͥ(e__Z%KЇ8xtv>zjo5wsiꮻtQ_[In{0 Sd4&v}Xț D?w{O[F5} ` Q۩-l?ɻ,\;o+;6f߸X6 ZC 4=Wy^o'fmi#n#of!_;w]TѵA0- -r7+K>]RAƝ߶ݥ˼a/c|ykR`F?lD[{9|)zԿ|>[a4,̓_z!ϼ0m%B>V7 Tw>t|5a_e0y8GBݝ$#\|[}'ʜ`zo*ܥ\*ez kM+WgS3E%/\Fsg.EFO4ˉq"s"?u4󏫜,5lWr|sy:| wfSu-˜}|$iظ20/aT>p+a+ K*YKuH%O*O*O\8+Q, $DXZ"0_#pGʅ@ceZBtYwyZ^A3(mZX0wWf0rD{&` }BuЇ0"Agk3h`A+#GL(!UB*͂ rXXUG+:\eS u,kU莆骥*.*P]jgw_y q֮,r7zjtU.`|Yb7i7G#إR {{1Ulw1an&C[[mZТ``Dn| lOw5Y  s%.ޜٰQ( C&iJwݽrl;i+&N[jnALmliuF5aR%yxt9(rcq I%'7iS}=1 l[$jE2CxKm7ҙ G[g6 ,W_Q5ŖΠfMx&j\wfqu;NPiaNBNJoRK1'3L W`?q߿'D#1)-!N๤lQ/* ̭0koBNhA ^g=O"daqx >F$&aIm=a ߩ72M﫮utAcl|3]jUWe0ֱul?}TbV6[a76}o%g 3ڂ8;LZz~8rJ~dm\ ajO4Ig00H;)0)QK%m?[f038ل@}$du,˹/r'Yʲ9y_;_Nn2Wel[Dx0RlXI)kI@J،s$y ĥjUVKwW %5+ΕUXqWk$-Iʅn]"wb;ss2MJ o?}vVN[S*YIUňgPQ%i-Moh7C׆) \KgkAB6:D(/`F,NZ/$[S$WhZr2'w Dlo7[*+( G4L{$vfn Eq$ NG6-@1VY ,`(3*h2:ڀmTYb0XJBԀQELf &v ^g]yn4L־wxg ]} 9E}q%(7s2V/SfvQެ}4w! 3l$?ZYH+?p߾k}ҏē\,#T %T"驱.xMG`xoL)U)w7`,69ҫvnpFKkےߜ)Y)>Z~).KsH)T 7aYڙqbó/_:2됮:$*8ʝ7\ 1qV{tE NVN57-a*!L/waጊKB0H=݅~oҖ{5hmigW~6$&mҸGi>~7&bj3K=t})w}M1Q*Pʖخ_B3z}/]$ -A')*_`m3f]ՖLܖ#]&B2o!^ϘO`zqE vL{\.qct99_A`B<U8[X1\99هƒXExwp>`4MQj .I$p) QsX%( %ň&s#w; ]CcZף'Lq/Xc}#J@q!a1I({:V>M RsMN1$ϊO19xHoJZ LfCjwtUbnLShyՍb(j QBo옎Fw7۠dAV_nAYo[+17ki&1Y-{ϯgQyb zDž: ٜ`8i*Y8f3P1fYe$׋"Wl.E38& \&ُYYx+ꆅGʳM͵-YE󋳭f'8tro`pw?L?8b;c1 Й#3ZĢOsЉϠm2xdr %sQ0齐&hMpY0Rfʈ]͜q%Ԯ6:.uC`wwVǨYiシ+b(2šX" 4G(UdD|2V3g?΂eoƋAZ㭈h*#iX-Um PؚHY. 胅F&tC(5VlzyZ^{1&iA^l]v-6t|zW+:^ Ǫ_tz& RW\!L@Y0"5( 'o eyk88%ZnN`PM@o1ȳ)g<rhab,}"fCG(c<71 筯P˜Tj>y'2+)\rCk*{F[O-EO6ZiccGgbGhpE8i~O Bv,M]rh%-FȥjLj*D=c& јTIĠw dYUfxt3$@ /d~Kw6 9\L{^‡7]oȝM·4-u#WKW2 pݳ/www7O9i,~pwA;K7{tghyT7 O|Ș?(yo=ixQwE|,aO+XAvpMV0YōqkQʼnLk'wNRRj"R_^i3U6k0Y'8l0G}([׭\e5uxaH 㞑ɶ1s1-P{] Fh2Wsb-५G+-4OmJAv7i'qq(@a;/:ׁy'"4iN]Ӝҧ 6*(BbFcT E8v& ,UQlU7QR&ިv>x79:͎)ޥL_ٟ˓`YUsn:HNc ʙZׁsBN|R"c%mfX0_w/]8?Vv~| GK⯚N:e b1vd_/כd_Ĥ"Z%hG1_) $JyKn \0Vr0.Ƞl1_F Aٮ[pm!yBKhaթNQ6#U!s0؁ C)1ŶnYO:ua  nș=y }e2)?6 EtЇYm^k.>8__޷>^w?\^weo\k>hno ;_K\cB<,\rwU{kuf[Td<oE_]w<9fs9FdVn+{']^37٤lRkvK${XcctOVs˓ӕi,CHmvRg"c3 eQx>YɌ}OZ.U"nBB)_}c>e*~ZLȶCR'A) ˗~AzNQBf<` QrMz1B'FfHztt>9wuzh!1sq-n0;5wIwGp:oVһcI3WM4ŅkzŁ'ٸ5} 5q;f3 ly#,ϒ,SAIk_w.[ꩨĀ<*"0^}V 9xBB&[AgueVԐA9zR tyUmm=.bO\9[U?*)Ifᵔ{*1> &2 F((J|CzGyj WKMOgKȒ**{p5s{BTra¡gn=5o%w: >F[&)xJ\;)DȬ ` p4LCs*λ`o R*il) z}j0΋,vM~377AZF6_-#;<@U@GV9%c>&'پ C:^zz{4#>R]` )6(#C1OL(hJ"ZVmlii%GAO_we_˔+S= 1joS'ϰ~wsqs۫~i+v#uA%jo>'@M!r[^^Y= omQ3Q=mA^t^jwHZv-7jEA'E3'U?mmWrU'~>9C7)2-/Չ=pcp#H\d#̉sCPymդw~_z4$EՁF{Y)H#̅PB"Ts$y,Xu!3Hj59EƐT>+^>D!'1B*i-p2=zc EU9֍i M㢔eQ|p1kXW yKt:~lۭEcr:;}o[+4kiV}[6S_}UT޿#˼qaucy6"e6)x@)|"@`}'ikr{Vb',yfF/K Τ@ǹ j#c5s#nr,62ފaaѢleUzvlr$,N+:7n&*GLtȌo2SB!(Ae2XJ`{!EMM5D#0L`9̂9 Jpy)]mt\ jڍ9":h2 g м  k]Yo#Ir+B?, ,`}X)qDQ݆#DRDf,)Z-bTU_dE|atI굖'$if C!,*5!b/iKB{H#8a cXJ?GT#VM%( |Cng$ Ӕ keDLrŬ"j8{TY/?/uI{ŴPh E]bw~tفQ<?Vyݹԋݲ>tv}g*ӂqY'3 HhǽJx$c*PǔqrQ[uggp\hz/fgit~,x8 oW=\c O^*֢h,$LMYR 3"IZ@& YXEZ" ^zbgaNMd;5:8opz9lnoZv' 1s.; ;=#!.՜|nzm;Ϳ5'eo2,_̭n%g+Q1-p\xFySbTAr=rG GE\=&YSj^Me`Fc:˄gZe,DHea1 G8d3aH٪d V; D":Lk\*<3]b>|-oP_YreF15ɶMSUQj"= WEYe;R|Uеgk1Vu&eGVY`%{Ckռ/VY֠ )[@V`̻(s@lć#q[V =Tk!vYGNI75#<&#T"&V&&V'#&^cj1 *,IZ*NT‚<_BT2qQ9ރf }Jw%S#3q5S;O !Tt97DGf4ZDj8# Y0J]4Ԃ ;#%6̗Mjۍ~]@"Vf=Atieb7?; F;Ǹej ,Q ņifu sfApep[JzOsup f,6@[S30z6t((A!l@\{ rbCݷ]/ MCpOhps lՋC3 gxF.&Xigv]Etл5Z_~ fUvsvY [> WY5.m=&<í?ZkZb1/ @-W\c0a0W2j̆UF1Tm./ *o*vr(%ԛBji wL ]+;3+x2:31scfc{f7\1,`&g#檔f&9<=f) u:PM+ڈġ;-qR`oe%{n=+(& ޲%I˅ j쥑TUe+[y>A6l=D7ؤ:.n-a~u|MJTRz]KTjo ɧ(LN@DU)+u_x RR+{ux[O33vWoPz^wFq;z]QJ/]Q[O$Z]_]eq9틻 y*KYՇtWE荻 wUUR>bs8p{ۯ; 󞤷< =|?nr\Ya ~}u0TW3L Nv7>پ`OW`/yG) UMP#tΕwiG)ʻ4Y먓Yxɀ48?PKhwH y\tQ@Y5NLqyv̝J[l;{E>e" O{l΀yNČӎ #YD})vVKOF#\*ʜ򭧜0_~t5?u^B\4'(sJH_MZ۹+.?]-}f(KR0Pj|&2턔RPB(Tc8¢AC':&s9:濔#2$m~WfSKu!>KU 0 0=rWY`{@\BD_UKwWYJQ}Hwe /&µvse?_w ,-PJ )44pd"ƅgrc]I.1O p4Z=̤dVahSCd%S `Ys k9M[Laj;F{MC6L ۆn4>܌?w}='% &h/$1 'I#HIJ%MG^Qs6F=x1HL2^%0 >H Q:A}sR~v &1RtErTjax8RE$uvsђL+A98uDom^7}ZOǫ{K7ok~ĔRmDWw2K/DȺ7 S qЎ"RY-] _N ڽ2DFsNk\/_hpBcTOmf>ͧ5]HS*yJ!w&\sH&GeZ@mXS;c:OZ#…Hh S눣VnBahWÇm]=º!rJjp[?]n}t7zxm βB3A*ÌgT\G}H*z8j"=yUUQN]*Z]p3еl+:RXc;CZ>>LǬa C.qp$n @՚9( ~:Bc j&> t\Rc;tTPgeL¯VqNLH'\/R.:Sz⳱?-hrh#^AJ O(6L3#S3\ 3/3PB{kPE`viv)̙Rj =+ٛ., bs]YJV>b+q.6$u?[ DII6 Bdi\`b>jGxW- MCC4|b%mՋC3hxF.&Xigv]Et{7"׍u\߳w94ypභ4N$mik|ss{>~K4_VEMg-o٩^ d@3:\15X>+fF A"x;o$.7eej0g+u vu}fêFqbx6J7z;9M!4l| dh[e-z}x5.>V/wI^ne zuv;+걺쒚c׋4[ z`fcfdBna٨q5ݝ\Ɠיdڭϥ;k~;;bhj9<-_4dz/lf) u:PM+ڈġ;-qR`oe)<+(& ޲%I˅ j쥑TUe+[y>A6=D7ؤ:.n-a~m'tMJTRz];$%3^rUuwW>&Y$i,5c"I0U[4"cĩĩHV::`hާ:wh3cN aEoe;\0#gU9~Y|j,лҽ[WxYٻ6r$W|]`a7a63{`2W[cYJr[/_Z%; nlVyX*i0Lt-V8FZ|Ny[Q9z'}Q?MXVoyolNv ^K}p͎ id4H fCooQ5;p]s@ixۨJl奷Ud:ܞX/P@;PKkR2D'"بBFt:gDYDZCpw=^Ay3~k3+vH}}}=(бArv}_̇mmwPoGPjSW8nPCͦ:h']MBKKs2.0YLe\Vw2m]ARdFw19 $7 )A ѣךklVtX.57;`<*7 +Nk42G8 i(Db'*AABFebdDD خFda(E΄\b6AW5Qg{&*YЦ"mb0}ڭ͎}ڪ];sEXä6;&E V 0Yy˕*`!sݪ=Rif^2$DdkbbAMPF,# Dᩖakl5_ø7P[E\75DZ̈ȬIk#(Zdf,H8"j-ĔFI5\쌏>L*)"IsiPZEIW.NCiYPh[]\70̙Y;ash)\\! I2)R.>]{{hfmȇ5Uw_|"* M8W.+y4n:+~Z e' ɰV&`.E~ep"_+YO{ 2kҸ*h-E$ ,ĬqEp:r{]jܻk~@ԼZvV;l^hiwb.'dlvG㾆d2J ^n%Lk#M`Z fuDDPBEM\,wVEv\L28l&*!˒9A GJkKҚ85N 2!GinѸb飛Ox34qX~ʻ9Ōa6YY Ѷ,Wt8+y9KAxCE:ft,Kg@e2yjevpZa8/ 歏'VN/ xa ZȈVs%dN(mzJ穀k&$]AL Rk©d 95  ol@Ox!dR1qi5-(hYluԚ8[<ܼ+_$}⮣H#?]Cj˛=&]W;S7)/h]6U>A(IEC4~3順z)~ u2* .&l Koi&VVk-E ta6eF'HhF9y^w_}'M4,a__kgʊVm0!bvRXs,%|*9ncpN u6`7L0tKs-M(F`0sv&Cb6 RJr. /!a"i8JuH*CsZ{ׂs[Cɚ F~ ~9rE=\s{?AmЅOVc9NKI6N9Yޟ|̻pZn*09w!Lzys{eq{ʤ[\K[Yo-\-,&xhx㶏5.*XˑKz8Ur!+vnÅqF߭ v{c9gd2 Ed10+K !/c79i4:I9ۤ.J6x\&e#ifQB"2N;)CfA]:Rܻ# KXժ},v:!2`W{;Ӌp23A/?f3y/ѽ7iۃ%flx OVrky4!6TROs BF̕uTLV>WyL;wmMa'u1ʀe ֛uDB!{HGL;(9 MN9a @`裱&I#6*fC6*ZAr%fZw WkdlcTtcM"9̪^a_Y{E6Vvlfؑ9eU|)Dx%+CFNl*vi vĸl$AC1#!' 2 )ee?Vxڱ8,@aysmp8_f74=_odiWBm2 `TMVOĠ!]^< p=Ѭ;+sXuf ﭖn|+wW\?JWׇb\=6s~sR! |~1OJ!\=tꡯPH|IVLsU쥘"n:s +>Xi\{1檈k2xi<>"A\};Jp[d`-Fಝ~G >RWss=ݣX*1G+oZuJp DiJ/DGUEM<4d).ίWYrRFkE 줢d '9#)~βL,I ] r#)I0%U)%5&yJI K9DO ubN^gpLĠ1 Eu,BbSЇd+8AbA\hE&f%^y9{\cSF=^iCsI$hvzDu;q]qW~A\خu7WHxsIZ\RA+MzsF=YwCZ,~hkJH- XM6R:(!x_exjU0;Իԟ{Vko3J~B#ʸn1zߊjwU=XnY|*A Oٻ6$u\b!S"eeY9~!ERR/if5=UտJw.C3v ߁*p&sEäՏ?W]B}YMajzpcrSk>ؗS&/jMЙet.Iʥd*Ee!*R[0EE ]*ׁ2[ ?WZeft-.]l[Ny35ݛQⰲ0=0,X/sd0fla%iCbb]UR#wVF9 qWX@I<ʉQ4DDaW#pGʅ嶍)dBn2H+(3oDyEBhTf"B=A"۪F3y7:q"3hcĠH)ű0BHYm+ ~kN5/JXz?ւӯJCJV/ q4ZYK0`$W"U#p TQ͍QmQ'(k#gl-ұ ж&̞LNN^4R:,'M_"WM4Pˉ8HD%o^Hi&UdWI %%߂9R`Zh|0+Lo5U3=q7Hł[٣K5P]4]Nz)޶w&u-ϠeSy>]}{8i4TQg.fFrA2RxF#Bk# QyP̌NȥNI R?G`fP nXf=ms(v.|7o8B7 & ɥb|P2㺸W"GX0taE$XLjAΫ𘱎P@G:&&VQRlzƌƁZ 1&+jN>?sS[#`iڪ 4k^vp/>}yc~1wP6ힰ·PoOPpbPu10*jeqoUh'YHUg6<q Zڽ11R^}0 x$ )p _ S띱Vc&yeĀhj4BZ"X9GߖfMyFmTd!>SJ*1!@@m2@S$ I E.(qCvYI(J1 {+%R.?9E4\ b0қarKLV.>4gxSr4 /W%I׷p :`.מz k7lg<,Yqj{ZF`W8)H6I,hpdӰ )0O BU$GkQr6"S Q[M*Ffd4sf"(ܚ195j]36ՅՅG g&pÆf5k/ݨW6OAaz][^6R2ţׁ0F04 o˜sG&2갷dϫ4$x6N0+D{CJGt`ZD"rR.2~Ƹ<;EkVkv%{c0R!b 4KiЂblt, QxE"ȅ4*bf9 Cv4H5> 'H >rTf}2'!jI1FjDY#V#qkhՔEYpb%XkϢTa$F8@sjD #!!Ɓ3uS/IXobA% 9 H\ˬ8ڂ^,'CיK6Ջ:^T^l*b1", #Î5USD<"@9AzqzPa68}7Ӈ[Pakas|UkZoQlbOk!ISf]ς<C3 OOw.C\ LQX&`z"*(Bf +‘"ih"hY=vpq싖ij+qdHP$ D[G' QFe,Ƞf2ta=rZ!Tjeqzx܎.dOgv^x?vg*⃊ J#1IigC,Ed\p*tv~|tԝ^΍I=K0'Utt}s:{ cz@i= 9ѣqt7|h֊t2cx*aKg֝qf?+w;nGl 1I=%|c{q9[vQvTexi{PH0M#1|aH0L43,A}*3f1~~2ØhQ Z=!FmzV~2mt9C-ANwBu*GT0/_nK'ыO<t ݴ%&|s4^W6D򃻺Gڵ*S9~n|7 $ϟç?uo?kཌK0H`$"8~Ϗ19n{nj/zNB//ꖻo{yJzO>?4xN+vϤ͓]+}5-Tltbq<aS1O`޻gFDAX;#w9LRe9)[OͥLJkLԺP0kٮ}l qOc͖jHecKHjNrfEMOL_U|)sf@OīMdڶĭ_&uh5 #郹=5M)|:u4&Y_&1<|՛H[6͝2VYׯ:6|_we`{j{|,jia 4LL)KEeN)7|2nީ;%*DJ[SK=[Ū%o\XΗ~9ݒo~gr˅V ҄W'O)*Ob\m5Dž[S૳R1Vʽ,(>zŋpW`RVm(\(˝.v7\x6yx77p~.5-$Z^f'cz~.' N^M\:5)!4r}/[b$Ғ;E؏&Bm,@FGj`)`CX͠"IVPVM)cmVmSTQ8? RQ:Ɏ2聀TM"5%Rp'_d{g$~ ^9FFfpYy.aFYr}Owr-/~4Z5ֵB>qPfmNuSmu&)#MBKvȑbC­*f]ZE.f$b]EV $Ȏ-Nُ|;[y{:_.UAP75JOsiWa Mk'|1jJ 3̍t&~v`e0tOm:zhvqκ1ժ<}k;;黏\c~bKzmRtۻvݟ|RWCԣOG\7' xmV::p EЯO:tn|a AMW޼꧓egU@m)@ͳUFoYJ7Ȇ@?皜.ҽsHLDT&O)>yܫyD}^2$Yfs9gXѰ2Xa?2"j[YJ.o}u-E݂<1>eGf!&nwȿSKLL[^w6rv< GΫMBjUMWDK!ن>_щq+N:(R\UaR_Kq|7p~{潋f\h凛&X OcK:հ4q͆J^Ez/iXO\Ή̑yY]r [ʟ|R6s}^@;{;'ې<%?6 ohKq|4eE|'Gr( 1o{G/tdэYypns7QJ_#uy11:;/G*lݎ9 âkNyf|}{sy\z#/T59Qb+֗ϔf 8ĎquCk ~DŽMݲ`Aޮ|>Rϗrg$}Mb/vs0%e!WCD)F$gV+ۍ T mV"!2$RJnV-`cg㜳WuQѭiNk9Ek\dK3xp-n"@| J5Kh-|b5`*`R'i!=%$B{>$3;-PQ{-`tlYYMkP? DE0C@/Z%:ȶ4\<lHƞGC*øl)C1<h@#r=㙽;[)$(0a `DI@~jNlbYS!ڽ?6>O."6%Ģ6#+Ⱦ1 ^{@#CˈzviC.IRXTتPL䡻%_!1ҡk@)@T2"TrdQ痝vSyx*ZvWq`-Pڀ@xxXBҢ V.B*g@BXx/dVO 8%ƇJF$\bQ1k.Jy5]0 XZձ%IqEA`yĢ2&JkBK*?#t=+}"7v$ ޹6 _K m`?l&wA6 >mel+biY-2m;jf*W3W ~8QP +sx`Oqt9/|ߗ|v>&oHjoL5 i~]X'RԳ t٬M0he'gBi.59Jk8[n~$ŧ9Ye8opyڎd}w:5l34=,\ɠ6N-(Y.fvז9elof&NL/71,k'Ngˍ!݅cHzXvRQZRk}PYi!ʜ_\h1MP[GIi$}Y/4IE$M]&4"Ԅ#Ȇ3[>Vh];wy=|\fSl9+M ~&G[:hVj^Z &13n(Fxvöb]QC;APks54O/:YؖtEE D9VAt$ IH:@t$ IH:@t$ IH:@t$ IH:@t$ IH:@t$ 5+aGO:5hGO:Vj$Iҁ`>vWkV}D g@Y Ⱦ+jZPeN(S'.n5-up;A.%8i# Z2e%xY$&^KXU'j 7ZsvOt^̨fVt5?p{#g'a!Y ةI\]0ƶ,l@}DIkQ;Ik"zk;j*p<ڛ+piG.s GQ80i2g$ 6X$ 1*kfW՚[2Uj#PYP*YhQyMьO77iz5?^8L.ݝ?-ewKl+3Ą̈́7J)HP*T9IP&V=h29֕,"#{Y\PBEF#f&lVr\,5gA\11EjW"Jm`#:a!Rœ VyLЙ[|Vc/>UaJi%hTgP4ȚHP,1( c%JҘ,5g6i[Cw*  O`Jb.TE čJ)(eܰŸN$)OZDCNkHu'!djI'J>L_(?s {bf?oqb n_zM[Fqk%j>&5Y/tP؊:a@׾p˫tYUctW`'-i %?z'w,uS0 Ea8+n~u> @6N&$beIeЖs~-߶QJ0x\{FӜxM"FfڿS4e |4P*P4{wd. vvGBU'r`\q 4^k9r)R@3%)Z(ƥC*P4>dv1q2^wPnC8$F=t!Go<ެkVOuK/4v]|+8`q]En3'_8̅ǧXqgiϞ,{lv1,} _CnS&G?٦κ$RSZ(3W笃&jEM4iJ> N%CeTPD5Sjn=i] ƿQf lWgzu G0nf|3(^jY>:{îC;GԎ^H;jibյiY{u kiH` 7"(X4+[n ?Oal%h<+x/ՖDE656%w>b(,N𛒪a {EͰZs +sy3upz _uwFnmx>vKvE^u@3oEt/iCw;*]{J6Pr&wjuzwW> ݣ敒h8R33懃ߑ2sGVSo9] ϋ{\ڹ5,v\~)O̗冶 [6oftk&Mp?V 'Q$f#$M4׀R,+gh~Fzh-Uw\zG'f\h,OYz*J¥U0ǚ9k5g㛩\>&]f/2BZy5ەXlT RҋjG};`>zW/*1j-7*ꫨ*ꫨzRY$q"IHD'8ĉ3$qb=8ĉIHD'8ĉ$N$q"IHD'8ĉ$N$q":$q"IHD'8jTJZ$q"IHD'8ĉ$NԎfڑj$N ĠI D'8ĉ$N$q"IHD'86\y8SrYNXQlhhYcMRZ%s>?R\Ro BD/3XqiZ0`Ʒ2\6M ܺIҗcUYd% 5Ƅ̼i Ϭ}^9s Y-ɰU9hp X 57Q(D "쉒3B}__fS1* b过|P2# uvhk1b{Y_xr҂il{'pP#!SNjy9D ә 9c&yMb$N2:QKxQ% b"FɘB}(ΰM3),CqT9* Y@/U9ەxƺڍ-{ʜcS">_smnۿ][$C$ݣgbچŴEYWoڞf 4x,ߵ>]~**tklA{U$ueM&a!O'003~yeP)hpɤ ^)MxsC _~ 1J/Ow6xżyyN~}<'_M~,<"Zrk!'a}Zc$%6@<]Y[1'AS룉VJB<⿙@9yEl9%qJFDd_տ @=7f)!U ǭ~ZxQ[7/xnk <|)g$tGhogLc:. 2hTb"Jk,kO Ǘ!KCj?#a&8 FXYRa54 “Ǫ~OsqE8Z]pe/8W~4+-6}Bxx+Q^ ޹:P'6J@fVDv;*/>dž1҃7,QgOdV*БcoU::fx:@(WllzK{㯛^hŪ7Ȁدx]w!6謪Anp]~27۾UN_יwxbz^Z igWLdT!3 6ҎR=* "0_Ne9kdmlHum G Rfw=_ZU_XVY:ҵD𙑮yc1'bZJ) GcZ2W,ap\Ⰶ^_\XA Ӛ#"vֱ$IOt)X-ϋJ+{p^c4D‰"(I<7(dZ4ątuTSYFS P}[ۥo3N>qQɓ`EMɓ\CKl-'_$+,yb|}ƔL3Fe?iJ()Q}i-㤖J. c%=˭pRb|Qw/)}P8ZǢQQѳ!^N쇣g^~!n*!h8HKU6{KC21)#k3)y7z ou(!ǫ䙋K{مOMmRrҰ3BX.چ f՟5XOMF \j.9`i^)M)JX mH֞SV)f >.f&zAR* A42kvd,Uaa1 e g-}df&!dwkncG_F~w#U &28-5 #M @`ɤP pg#9YM.$LP };Mb`"|L+]:Ts#Üy*]L:Ne(ZF&Bi¬F FrKtId`^k|b(&!N T3'R@/kB$^ȴHc.D Iѩȧ alƩx2XJ>EDUU"nxc4:C"E8IB+I 4W<'//\TqIRQq5D%U8=\Ap]Gv\~ѱ-qqc4>ߕl"hF'vkN1M'6؅A0%N' rAjn1m +qi H#J:ުCٌ$G)%빕-jMg 0Fhq_N٩8TyRzoYǙsD &\ (h%Xt 9@\'PvH *WFxI訹YRT)ƄܢR%OJC)֜gYd8ӚsV>\WtyLKsu`ع'\Raй$r>}7? .L4B1}WJhCū6F KGߜU?Lr,ٯ<';KSf[ITo>8p/?VgRvlmn/CΊH0jP(.W0>yp3l 1WO۠;m`hЋZi4bM͹Hry)<1ԉ9tmopɃ0Ȅ>ʋ9Lۥ ۾|dPGH#ow)|Uy n,2mmW;hєs+2{~;܃_#$zKaz;Vb2ٌTe?ZsY4٤4]Լ%6nL\Tefw7keLӕެ/Ιٳ_*G[\}o>L]8'̬A&jy*eN/NJ'`Z^ -8d ׉\ӣ^Cyr.o-E"\Ԇ(I̥#K +FqRKH!hf(řP 82Q1xΒj}ʐŚ=1qaxJXl~نHm<)~zOBb@O`+}~|9WuÚkmƲE0e0.lcggLmeIv{YUʲdJvUıX\ԯXpZOvqQ(i0~# .yhK>)*\\\]AA `0$2S59l1. yQ,"-iZL)nߊGk|!y󊚿B念 kHޗ{@{ɓmqD1<8.p`3}$tQz*wW!fV$(HWe7J`'W"^X;c$zgW~.P\2w>T[Gquv,IyMp7]>cC*~Łtt8ƣЈ+edgvsʁ^LHJI3`aJ=\bA}L&W>Tra^]F vsY:pVm^QA{@݅^ticSU̳: {ᢒw2;uW'nROQF YL8t;})Ll@+lP֙ghbsn^y{0\BW M+@I *X6+ k ]%<\]%tPN]#]I2c0t= -m<]%\utut0˜`ٞWF]ZOW %Jcn]imUQt*tю;V=۫sNߘ+W`?ւj/f݃hGWV=Zq"!ZCWK©oZvtutEň`᪶UB+Jթs+JmRW .mV6^]%Mի#Zs"bW88.L>xҎΐ8Udk JutJ(+::iSj]Z]%TڷվJ9ҕdH6+o ]%5`BKq*d9ҕbio]`c &o} O'LututĬMD`՞vKPk]%6J:fXq]X>p1zk7wJްt::1s0ׄUx[*tPj!"Etp m7^]J;:G*њ ;ʘ8]ر_fN^K\رW$mʁx,|/]DʋHeEZn{U ]uŽ}* Urk.CK0KXLlhR;&vB)Egb̈́R~*eYoMhh:]%vtut!eN@]w8iX,ڳ"U-S9^$%g)J+3EevkFijJ|en,ҚE\ :̣xǏfR3o)wg2V0B7 jф@CHT'+RyDh߃$pe /\?9PLw2_PQcRgʟwWO;"g]_ՒM?~ޅŻ-񭯮%0(g7\\5mfO "%α uE*GBY+^bED* MV(ENJa 9ꨲ{#1aHmyg/U8dMǥp? `L",DDꥦ0 c<)c"OTrpR{LRLYhj@1-D(+ C[R|͖)ލF ʏsû?)j3M>";En{ 8Õ=|MIL>fjƅlc6h*ֿJM}lcOxx 4)E|b+0`܆`&>BzPZ~396* *R( !@T ɴqNk1I!w LaYa4Q&q'N"X4Q k R+1`"@SL sd:/zc xbHn4 y/Qo :ocA8׽/T^/| il:?1zf89zg:?􍉳4q#'g, f=%Z݈8h]|܋׸ [.oiіO3q,?*N5duɀ1T)w68R$#Z{1𚃎F5ImXO\'ҫM{`sB7'0 \ }ȵek(Q}SuϣhJؿ+l]tw]`TM^CnÚB-TEA:FsPI7T$Yo ~NG}:jmCZ|^YU:թ*f׉9N[&mHiSZdjf`2K-,!KpojX^r+lK(O+; (%9HCܺdѦrs^bQ$QN%j!" p:ZHֺI9DŽRn ]<[֑aY ;X+\$JH!uGgz* 5iCg I7>%Q[M*Ffd93Hnf vfOƅCPt\*.\ Jy%e@e~f ͅ_p0do La"b$aV !h@1G8L(`aoAD_/cSE#bHƞQk %6N0+D{CJGt`ZD"rRս.8ێGbb6kkMamѱv`_l6LiTR7hAB6:D( RFekҸ32*hA),>%,g!fdbI64AoEw{5Z=j|Cf -fXQWJO}NFd sLk#m`: vudK=$|'r*\\.: 8PِdqA#ʁ5}9sz.3Au^1{}6()Fr,_ɲ H^ჟV7`W,)_ ٕ*U3_$ckWV9Grrky2h<)ZY91铱LKRAȥ-\3r$%zsD@MZ9k|mn(>4]g%dWia@S' YjFڠZ~Ύ3-6¼c3b<4֒&ۃY؅:G跭}J F18bؿHfR>_VFzP4~P :2',jU+ V^z[e!Khz: 4@LiʍPKkRZڨ"DDgyo#D<3DA˸`!ue[@kTqD%|¬l&x(=nQYW whl'|꽒ZAZ蒜?=^nw\=&" TW*%jµPm4FCVZy;T$aXu37eRKT$2dtGt((NY)Xn*˳ "g:#/&,F?^~FjpY@iκcO[XT`Fꫯ*W'Vը(Vr*;^ye:˕ͩtaUgҷAN=|_U^k{*\ K;uN<30A JBk 96BcY21do͹.H!)4c`vcƟziQ|׃n1+6 ˨`eα`w,T9H v6s#Hnlfڣi[G:\2` K 0fh@nYFzNՔB,rͺ)HETw_S>B۵B QH/7qjɚ+>9Mם~+4H$I:_̻4*0r!}}T|0οG\&OjS&5y?#;wl.rC@qN1&W{ua5 nQ,9ƒw`DH6ar clJn9'3wzRvQ]9W匿c;KkOt۹r jItt4?䜬XZѐMM.(tj gddJ;œƭ;Hu|{s}6ͼr1Nt{l`gk;(.yszŽHcdlHLKGBHˆˇAt1^QeI]Nz8zQ[=dF]6WΖtXn4҄G?Pʝ?]鏮<Tَ08u;?|;};)+Uċ+~&KF8]/v+Zrx>q*SUSw4q'_ߝ|滷ߗ}{~{{N0'oo,JL2FѾ/" 7?MSCxC+lr_c\͚qqw\a!t)ߗ۸cyXi#-jۋuAVPilj2zz5?Yv1UF/H6N&"YY M@mK7<߿;Q?9F]*6x\ 40Y0Z($Cix<3 iaUQmDV5 E"*pR0zElJFϜ%Xj,Qm/뛫])'؞xV9Yd9iIxUԎZMoZݙt z@l^'J+RAS~PIs݆J@hR[se1"OU^D#Ym׭u㤳et4LuD#2gSi z[0;ph,q3QпȄQ1! ٨h皴2i|0MFh͘ٶˣ+jX[ʘ{XRv4]B͇-͢\(jG&jG+d3<KEEIE5 7,Ď]z.(_8>>&?'oM8~JFqӆK+a߽Z7K$eTiL1 ( ÚnxZ:8Vߌ1A6ё擫Dz.cێuzߣE.7W9=N ci7QGkFg/udɋ<;7p$|hΛj?݌ ;^u2O_ڍ`Jjw^NJ_U5  k͗zAT20(]qvis].)l$YdztF f,E\s1]F,O5duA;OJfa۔G>刉b:sȑ΂#Qi/TÆlcy5F lƁ(]0bˎGw݌ȕ Kr}}Q]b D ]ԺKjzu#NsN%֜Lt%5oοY5. bY[n+u\))hZ (v(liX)%*`BHo>o]E],sFe&K0C(8Ɍ ,Ԩ}Rr ]RdYJ_.l 5Ƅ,Ghq#q$t ?LʢP\Uqo[2ժ[P/i[Ϫ-*J*&7f?{seY]]^4Hb4M :U7*:xKfM! |*fZ;(+W)䜑?iRβL̞@0bt†D_~^}8a'!˃ xX=@\=c_(EM2OTԘ)%9rd>a&s:sǠQ&rZ8r}<79\ 7ߎnkNmP]nimW] {iM#DZrؗHB\VFZ*miDZ5_2ׯ.F,DuӨYL\=JvK\'+يznG⊗q7⪐ l_UvT;sQ) W GF{# roUV]WJmZq+p|B.}WZ%w]\*q[qň+p#qUboU!ט}WD|Ux+@qZ"W׮4j_fiTZՊ/Q\I+j0]k Qޏ+Vw`X%= >(gdz:0pY4$4JpIBRqlvBteMʉhP2Sw#/uizOvdn{Btјj0Os@B`ivm@k'߹kGO18[$j,m̜ImƘ!uK(Q,,U SH5gm͙,j̺V ˨`eED%r18)@+@[*Hf_{4cc(-#5Hk. )ׄ+H[ck!f"ֵxO\I,t)A"I* liTaxB?#;wiab2J:ъj!tp.mwxMB.l5~6iKig`=wUL0Nr9>/r!v\-HƫD+{<<N%$dXJ:{7Պ0;G][s[9r+zLRul\~;TvT7IcԐeϯ_"# HTٲL4O=>B*o!qӚ_/6iQb^Ύ6/7j§upKb JK\AQj_h.1{[[۶fD{3ϝͬOhd"/rm=zW?Ytۤ xnܨUKxt6V}t3n1Ya1UY$9^{N dDUb* ˬc ts"&qB#6NȰZwz^dm)|1(O.[e9M7~||[kxG~ۙY.^vz|w߭-ܫ&ήMs`4wTPwեv!w#`n\zt3/]}>92U;GsC~1[W5Oj^_(Gk%d=ϖ `586^۱ag0L=L0u37«ZiF9H-!7*M*@U I eEc&jU:@;k􎒴 Aqz2f  i%ހ Ipcf<ŪPU*T%R ,)劇&S̊sfs&F&SmMjUn6,;jdq:msq&~Kǀ33^M1@7Ơ1[w+Wʩlnp|74I{t=UX$;y妔 /h2BLH`܇@}U*@lL4Ȓ:܀J̀8ČA B"!؂ӞYİ2֝MG}'uIq3U^jgz{Cq{$&-I8YFɴ%21F[>AʨRudE $y <Xl HЊD^#H|mY;#Z$XwXÔc:^+.d[g)I&]Lz촃nrQ[F`.GylI ,2o.&rts`o$bJxmU\.̿} 3͸ppG\!q \NϿ}F-iJw%|̐m),|TEx WqB'lym݈\JH62(U0lRQ?TTʞz*K3( 7`P*y4LRuRS)_>hΉek )K 8frfq}C!@'/3lM8ifjDE3 k9Ѱ: k 9VOsLh>AA*,D&ay*9:iʿnP;N:% 2G|fg]D4SyejH{F'as, yn5"2#ZφRd6!D C_js)]l2$P>.owBw0'RNmb.udLw6?T]g8uNJj Q+h]t '炡"xDŠ@zzDlSzqs#r =::|}zph ȴd-qDdGD,%oE`^xE8E(syc0 a8a,'1ZpHkB ˯)*oEH4s`*!OB< #)Y}B cIJN3W=$pƗݵlG._HGEeey8񘫈)0Gs)ӑc3 dJf5!(1Hs`  EhIt0E&t^2˂p\ :D;nT97}8*\{ףO/V(.`'eby/ZI+ TkBUCѫb 鲙/w~=F676Yc@'?w9r鸊mpZEQoluߒlAtV$^_ơ÷)M+pyxk hlj'(8%JÃ* DHݜbF(;x,߈YZDN|juIAFSL+I6N j4g߲|N:N:N:N/ݨFLTQ}LheY,BR:#$pEj6smmܒT&pyhWtvUظn*>8۹xwQ&2i$ÍLf&3EpȤ o p!OВ $Lðx,z{LR,fl~hc_xdn$݁o~6dK<Y]sF?yYKiYP;N,2?t|1yr/_jԔ<(l.g}Y\FVo"Ԁg8IB A83eލDӉuud(j0 mR@ruG0@yM)\2D>_.eީtƽM<{~6nOE |1lT?07bhǜPC EΧMwshEl"@b#Ay)21]9;5WEp|Q_F$*R*ad^J'DI8s>\1H1jeY^Bgp[$D. MhrJP*Ft4'tױX'6L#pYk s{mL\%32,DɽN 53Վwo!<wm-p[|&ݶ@t_Z|ؒ+Mѫ_dֱ-%L$N! 93'%\Ha ̄3q6+jj !5#2KӨpr-v~7]fʎ31]dVd]5=&7Mi[j":B)*o`#ޢnT4i69K6y$̘~ס[*et1{ʉ7^@ifGNQ@R]#cglFlΰP,T  M.e& ge{ 88n;=~:'_8b;hG2H,dP%!HȨL|0 zdri"g{.xMtPFmԙ|;flhZC,h1bw&fA\q1;[ڪv`\9΢uhaRƝ VE V 0Yy˕*`!s)f HR3/2$/: kbbAMPd8BƜYFȩNdOu[~"[qoPD#Xo)<3#,(@fM X1G) 0"3`G=1둥Hq(fAD<I%Z$yp03qȳ8I=x^x:;[%E1..|ct,YΝ 9[aG.s{$2)R}C0<|P~Z#75! i(F6.Ֆя7>{^6%ey'-Ir=T{rL >*ru(pUpUTWW,l ;\)pU9bb?-]nnPRqbhq+GϑzA.+[ l7Ya$cYfPǥf`@Cin{A(A|췤$@]Vߠӻ[iJ˟Hc} ZENR،i^CjB44.%\4i>GNmUNTkp6j {R,6VKl1d9:K+VP}ɏ( P!oYXsS2\#]ެf1u'kdd)*qQlݹE`).⪃9,>.vrXcu9juH$p0pUPHpERaWzJr-!YW$ñrHWJ0 \q9*pEX *J1q:P^;oˍ{n4.˒l~QʃƯӇ%Fy1k,r; H.6hn 8 Jcv9ShxzF/ሺa++8X-.D;/QovxўmF𢷒4&Ǖw%/O3֜ e0р92`CpIPp̓װCQ`{9?I"ﺂW˛z6Ww/~>w/޹we/8T7lWZH}m52"JJ2A N^ l)K.y¿}w ) wB˴^\;ˑYs-j ގ]Ϝ_Ӟ sf)lyB̼7m@\Pl/!z9 J%ZN+njBRV J[)t+nЭBSգkz%rRVGrRVNIY9)+'e夬_'_>8(Jĕ %HkDZԪL_c!T J[)t+nЭBRV J[)t+mdBRV,BRV J[QRV J[)t+"V J[)t+nJ[)t+nЭBRLZGNg }1r5ANhm"4=5HۘȂ,R : %|0Ϊ<,݇AʨZ4.BF %n}FrKfO b8"8L՜ 3Bar2c$jrxyrF 8Gnnvugl9.4=y|;َs< qs` )ufJ{g!6br2}fjS㯧 d J@>GZ3KP#'(*2mY`Afo $ЙdepLφ%'Zs.*Jgl[/Nh85V?}ޯxuW7׿]fGFJyVA#Aデ:b}9+e=PGTY/r}uiKIyOc#}tkIê/JB FʑK9Ay(8b\m2KuɔB.&;5~(Pġd/L*&.mߓ' 9ZXL-:yW$}]f~JVΨ}>e}^Yu;檦(y-W ^hrb.PZA{#e ' 7?I5K.s$W:}ЄӋq\svq+\ĵ"Ԇk :k02Xt9sI Y4@\Zt/> nKLVXHGʊϭ,`B2jϥ2X H,TrHAU BpBX.YʡnMO5mHZDE.ǙL` Th@Tw l& htZWEV;å2xGlvzC0~~gq_=\sޖڢ_8=j/~'o5i~jk@FMږPQ W fm#7S&x? ]s&׸#0 )Ɓ`}Jȍ̟^'gG6ŒCZY9X.0KPԁinװidD+n xḬ4Lyb=OIˏwZ OZ$zG XQ)rIvD7J(yA{I/M)ݕweL{1#k͏o.&'ޮgB3?ȧ'ӹv9+E\LښoO`1i[MڞPٺnnPfu'Ei0l1fbɇD}'ݽuc rJ#ut,ic8jY.Ƅ_WN_ӛUl"fs i/ 8lylc{&^XBBlEeb/Ih8:V{a tI%*Z?~ R߾7|9~|C{LR(z׉|gvF?]] [tQ&|~5rk~Fb !l΅[R#˶c#]&umX@{ vs]}s$,u5 o_(Uy;g,.W Z\Φdt1@b9[Fq!mj&ٚxPN2iރg>*j#7k# ^bx牷RcZ\Ʒ &,lE̍u4L6>xL]&T vYWD0FВFɹ>g7QL!B(vPJu[ sGcM dFN#FL1`FE+=hRȥ]Jvu@BHE zݕ`Úܗ4xs6M;TW7n:z&lR-?RR}zgsTEZH9HT_)A: okiX \4n ٸMrV; ײ ¾](ف8ɒIə1R8r1.Y&XvHS]O} >l_i+m_vnhz otwԖB)x>9woG3>"1K肞nvft])wI0%U)+5&^x擰9r>a&s:cݒ,Ygz8_i7رR}3eaĘ"op(7 ERǑhY]SUUO]#dB}(ΐ'8,YַEf%G]uսe1)S] ,d!ZgEF]Zu;twTv~Kw5 ;X݄г\bK'Lk{ՅRF,AsebI(PzwI=@yRf'pMmVi,-`i<ћlv9ONîe 1QVb?X m?TŢ??wWqPLp\!XUiz[E~wyo{qe04}ʼ*yʐM=~0W=Y <ި UхB kRR1 e|+I䜻!ٶ)\Q/Go/j:ա(kuk t?љͶCZGBZ"/ٻd G%⪍4%,xVLkYJ𧓞80z RUS=wOl>AA2 pʨ`1 52ŗVU '. hm%u 3H3KZ{%fkq.aJH=Wo9T3 3NϦgfwWJs=3aqq6n}ӡ׻ϗ&!0ʃ!K-[bI1YWR yo t !1Fg:|Eh'һ,o&N]j \fW)%EtpQ#T)SxP~+2Jdh%~Я)yVQ I4v[#ݽgJ}ɫm lӉJ i&NUhW^Q9{~[|:c|"-;)lyYO?oI>|QVN&=9UɟspJ=N_SdyaxCCtLy;`XJwb$9r5⸅JwV0vBq9  P龷(\!Ú1W]8sUrng^BL9+dv6*-S7WK):s GV0%|.s1W[y 3W\UfX1W.\U̥\Bs%&̰:s}GO2W-'2\w+))~F 욳9Bn%n2ЙWiaY/DGT8%~X1&79y3 wM+G|ަuǻi[ r*{Jҿu}X ,ÛfbbPk("UX16N*_zUݣV+,ߨV&bvc~dW 7mM=vTP2.`<,hFcUr$C7o|7x)=Ez0.G4j=IƫDKm;Z/Hxv9ءKfuZsq\.Yu"|ҌkNW,LQdIj8<2΅H©I@e$gSGϔT[cgs8 Q'u%URcDe^2"0n;o|#܀zrKY.R#K!$r2W,AsebI(PzwI=@Y忶i*ܛ~*Kqn=iL1gYq!ߊ\6dZ5E$ .hCL$]TwItOZ#…ȬӮHګc,8W ܟ,]Ň*bf8(&8.whCjͼ-o"Wv782 dW|_d6]̬^}8fr){v/PslzԊmtzVA#A)MP//3[QIP&sJFƆTfgۦpQ+[ ڿeu(k#Z/vZ]~tfk~֑](g=tER~[IN>:sɺZ1Gsv|t8z׻ q\>8*?W-xA @5GYJ𧓞80GPBIQSOßAA2 pʨ`1 52ŗVU '.o%~ RȄ/d^xݝۑ׍_ɵmtf`孭;OH!`O\OBhECm 2NJ` (8T@_nD|拪T,2P.{p VS;Dg,R % oCq}h#a<$2bFXp\ ouhCW[3l;= V΅9 `s;_(6v:f<.y|^.GMAGtDp&tRkfTJI'ls+`edy| jSzi[^tu:t=c7Ck |`jm#L*ťɒ ""dY^> Cg0ZMवRz̈́' yEld$YɈJ ,2dtCKQPtH-CD윶˼NOhM!&1ym \TBYJ2Rc}.{;^AICiU ֈY7JyRv3%%eiZyu6&E;P팑)lkP&l(QIDV s,š9ѱKid)hWtQӅJy('Qb3k}h"pAup k! fED+K<`l1F>>\NވO(2KXo!p9g]HxRGQGHYAtÙB(,q&P(f]5ԟR2DEb9=k/پPg1y\o2q=>ЮcCuv/GuP2:|6򜜶v0-rնwukS}~,;E>"=Hg>ZyehbC$޷0 "jcል%o^z0cɑ%^L>'W za=F_Fَ*aag3Xhz,<*,\JT6 $N';BG_+Glo-2L%$1*@Ĩq) `*B ɔ-b+ r7T*dMel4 &+)N8*)#sѺw%vaZڝ͎CAmӣvn{)yt䄶NH"Y&hAī'BXL iB1IFLENhc} N2 5C@fQeXFfa1p‡TtIuf}cO>_~;pMv7JE _T ߦ9/Yt{7HS| |5y"K Q:-PS-TVbc;ubXSo =K>?!|Ge]*"_M?3l{,ѿ3WLR4 %am('0G3CL+^=iFӲƃѮ>3Sz^3b JDb2)k3 TdIG & uU,!GʷGe~|́S3Y[k,-JPAP|J'$+gJQxeC 2VG sUYvEdu&DֲP(ERhPG.(IYjfPRgH`%>Q dt5yR=--.=)I+ep58O$VlE,OOF_pqf$9SjR] KWMoM.'KwrL/cu>NjXx @5N @S?p~}:2^sjߝ#Ϭ)Bs TmuB;d?6Y+>mt0-{瑯0i-u>Kɼg[6/EM/.߭~H Xq5 ;gfڦWW|6oTvVo^fF5}WO>*g8 s|piַnvPSmմ .;<C!Ԥg 5]FjV'˯SZXwzѓ:JE#պ 4ttk>c-jUNEhvr_F̈V%*/W_Xo ]/|~%u+xN⟜pÃI ΥRg?gM gO/?_?ᇟ~_>~A?}{y\9ؗD@d$4.늢tUjgWbMAG@M?x_` Br?}r6I/OZ6k,$iX,1*ֆ;kW(F}n4,*R$ o4d: &fvm*?{^R*1>$PZX3bTZahC{Yϣ-FOwCU|M:@ G&G/Ϣ&c˟Ѥ=rK{V:9c:Y5̂9$*[UdBGE'zet⯘ØP}A7_#,'O'o~_*zXmvJ[l#klϣEk颥~U}a$"7\%t~ѦJd7;p͢q |BC^nM6g+^9_iqU𜩚ٮU|[Euvj9}8n6{ Nȍkԃ+BVzɃN!%\9UqSwiG OzAI%rOr|9t_D;>IPQ}7F2lF֮40|RJ # 7~nʖki֛[Fd(F(}t4*K)']i",Ngj[xV+zsEu$Z1[ǚ}(F\3 KEL}g^t>͑"QO !΢R[nG=xUʓdȭ_h,%^=@@-Q!*S@c$m+k*hM MoJ۠Q›`<{7&y5Pt: º)7?R' VWWNyۣ̌\2XfP2Vsxa~!qcx\֘" ,V쳯`¨2Gd= b>)E T &^Gy{Vv-5?{ KuM;&i,tdt0\rQ5ݡUY{ڳp\M]Mwv0x:saBDY~r_GP᳐|#@ HL9OJ\-/ycw/ڔOW#h?+ y{|·++8a8G[ wf ח[6|}qs?|kJX^S]f('r8EJEPibl ZHsڡuoQŝ70_jE4_CD,f2vu:Mٹ!\FDiLX%Λ:)#"D bFQC"eL>9W?.,n T2C-.5u8ZV{}[ha"[*<6|^Хѧ u]nK~+e3.ȳb6|".uiQ[սZ_Fâ3򪽜f+kvc$[,̕RN!>J%6i$ eG9M.$!b4UTs&qFpF6x5t\\ y^^`N0'/UŠ5)e}h)wYhMIFfXB;L;f8 ޵O!؝̥*<ד RR"΢. |6dg񙖌 r, +?_gyAIϢ/e2f~]5k nkUwL X|݅](xdk>>Jń`:5.fXD^S#ӽ'<2?t<U vK*f1L.!*!:M<gD(EtVJp*d* C߶ӎ90]|RZaV챍v쒙 薙Z~FX Ym dټ\)X홧 溱àJZ8=0st.k1JQYr!/A\$YJH$G}{fa5iQ3fmj+mnlCo%{ ,tW܈gƹR^"&3+IMBY1)VNa\J)jsm8=oOː}ndp `(^R: tNxˋQr/?SbZb3Oxxac%*Iɷѻt,r6yp^/>Zs].|O$[hDcH5{Тp Gy!a,ߊDN)E)sIy^*>D?8]+|-N<%24&?p_^jKy0aKmٺNv34}cI.=4͝ȧŁj["jh>DouMmQS12:*^6e#ɃJ l刮]:AJ 6֥u^#UԮRgԇPAjEh[մRFiefq.4kfW!>u;ףn ?`H|G8hUmTlr+˚;8M#|RP.|=S~dQS}< GBysа9^.Wʾf3WC*ėO4i =C(5J'Rzt(66ȼ]f5+d:*`^3$0Nd0`R{qNv0( Rb؏j\Dpʇ"֝vh*KDk"Ϯk>egttS_m\`"53%OJ K ":ՉH!G4KHR&WM+Kّ5RoQK/Z4g!-b LʄmMqI9֢.*e M3'06(d;4=BzAm)rrOuݷ]$oM7n-ƋcRwN6]Lnv£ɹ*gn&m&NY/9JXKɅPT;,LTѮD h jP*G{Q䕈m؜>Ys.`YϦaB[9IēQb-9Rp 3}U"n8.`$`c, xBE)G&^2?Dk'rYfS%v:OLWչ4Q_[Xͷ/t9th8(A`_;r'Qt9w/h|= U/їl7ʈi.|ɅBU<Ǣj<Zk<ȐĒq/NۂENQL)MbfqiM,aLv㰣V*[5گWoeVF+kFגe3moi"BI7K Ll<=LfAò ]4f?orf"JQ0P'&!1J3%jz/K~IwyP2j~(Ke{A_bQ$('"!j!"~4&rsL(kC)L)"e)xdqRʦ$I$#RHDӶOge:crMUZ[r7ooP?Ҥ5Ci4sg$؛>2}8a | zo~).n!bK-:>O-XCƠ/goĆXZW*ƆﵹJ()#D+k\N*j&ahրbV"~&+Z=c+jgWyҷfʴ4Ok?a}ys2Z'{T{<2w,sM {?,|xM^7pɽ)[ږ;r4STޟB@Gz2XJxmHmbev5%oVTcɑkfqA&-I@/ R]#cgGJgX8 u o%cx 䵽,^f'BGpv#7L.K0Zc)d @:CQA!mSlaɔ-bB< I ADMAS2%SضI.%"]t%~qZ31O;[ںG\{)dX!ΰ0&%RIN2ZC,tEic,(f@lEĞ&eF12s‡TT|c. v:Nqx't@G>cgFsFAG?>Q#gOVM>*.>|UySΧY4}>ͯ1&9c^ޅ<͋:'w~<׸e2Ydh^DC7H /3j{]'>j(j(YVI[t9>A!QG#t!Jc7 فb} d َĤ5$ B ؐ2I2IiIfom$t"\n90j{|lxpsB!GG ʆZy`E@+ufh}u>砚_ouY-2hAGK˜P/`ޙW@N5 u5с*4hTֻ2bYJ`*jT%|;g˳NI2Lʖճg_U9)vE{Gu`ăe|P^響\1 ڲ] (}AH_1|ޮD/ީN{-e9ЇIXP!['"׼xOHvS51qZBˊo2a]|ؼn&nC yF[lQy|G?njf冋mvOӋ] ] S$WUW";;;c1O,mO:o7v;v3{w<ٞFoMGð{6͆\^9]Q˫7;7wz݇֫148c=\۵QWb\8i6;jчlIb[˭͟6WH֖2yy;dȽy~67ߢ"|5Dljp5oFZ'H?Y#]bv9>xT zИJsQW!P9k%52 lv@w!'BV{))rwS |-֮o/p 8\lZwTmy$}Y1i;~,{4uOILWuWVf~I̥ M!zSSNZ闄L鎔)krbwWW[u:S_=lMx>Y03[[̺cZaIHt$$ܒuq1/ɝ9il3/(]-5[sէX="o`KfF_||W_01e].'並V>}⡠׃;dj:!:@uN PȳKqhDxg|F7g+4q~z^7KbxfY~hE/WŏqZdg񠷸e M3޼IۡM׼Wה7+*ծ2cZ=^݌mɏo?xNȸ[a.Ḍosn5k"^j_֯vrGy '=mݐb1(:u'G7ˉ^ v NmUݽ.mcEPn5z:R;V|> G`2?Q-V+x8m\vbsL[ī+~%Ӂ\~i_BׅTH;y| /YIϟ}ëo?|i߼՛_¿rc-kM-?>K!G]׫_>kU߲kt@ךb׈.߿W:Xm"bCGq'򺍱FdILJxx~xa)lN5R$*V0\<.7}Cux"ؐkd#7;D¦U?OxM&YdvD}&Ⳃ?ze.gkJL ,(9C4Jr!T'P|u}y* :[G+`ݑ]+m79:PVw]C}[Y"%g&IQz:$UJm$_a~`|TQ6fRoJH [DqچH&d뛠IUKN}.ŧͥw2xV[0*8 M51HP0j1Nͺ { :C6$KZ؄)R,V'6|c3qslnїWtQ{j:)4M/aH1ws썹1PQ7}T v7Se 8^8h尡T)M9M^{ާZ1V8ҵN(9DT`W%Hm%ZyҀ1m;eHD#2eM( *E(E 6?h#t3qCFfBE`:$z/i?%m^֫ٛ䣧OD1kUsNe͊*m.G~C.ALCJB-td РrD'==UZDJ%02wI _!XJ!VlxiCYSžNf_'s+|7lke.|=uܬ?|{x%_b8U(D5_FW,9+ъ\bb N*^`J|e]`yD +T ̉noN7mP=nm']FQ LSm7F?ȶG#E<{mgF݀F8L_Hᅪ/6M4(Aֹ+X2!J_9 2Bͳfuʘق@+iv̶>'3ClkR@ܨc9ɇ-[ݒ֢v-SI#s+`;4v+LlCa-{z]bDǒl3ӄabΧ`g>ܬ-Ń1k1L(,Xudӡ}5^t^D`A֙/+~&6HρML!rO+z3l1ٳ;<m>1h-OMhyőQ6>Z J *ʡ.C,d,C^m {AicmMshX,[ɺ+V`6πkK4/Dj;ሦotg&1FJ*5T@=c޽[8:2kX Ŝ`' (WJ*-j@F.'9ʹ`Xp BHd`8X`OJ+{X;N͐jPo]i?@6@0vO{? k<v]`&18ѠZ'€:cJC,p q(hH;,'ka /s~ ڹK Q&g}ˮdI>St>S Uf&5,:J gyahWu^ 4M6LN輮a>DO:#.Ks/~πyILPc $ku <oĝ 1ԧ],ԀGwf$Ywl@F(`]0rpc˻ǟ]b96fiX} ][{ #C+~uic?bWoFeb ^sM@f'| $$zbY84+tOlO`E_qh"Rܧ% F8k;1Z)6TyǴaxD 5x 30|׋\'Xנ 8EژZ'c6# #mÆ[ Wjvzr3ΰ %ߒr 8oT[!@׾hϏl6ӥd) E^qgAPcv cTP kbσ51z7%a2kV3_,N čvXLJKFdx]kRWIۥ) D,0rj;VAj]m C|9O]X~ H`y/(@vţB`﨧5X`j?.K>ż-ny}q0' x"ܩOL>͇Os_*0ND5d?5,5&Cspo/GzҾ6[NE՘y:sX( ֹSDJHЦp$A0\$Pv~kH@$&4 I MiH@$&4 I MiH@$&4 I MiH@$&4 I MiH@$&4 I MiH@$&4 &H(r~$e0Zx'yM]cH6&4 I MiH@$&4 I MiH@$&4 I MiH@$&4 I MiH@$&4 I MiH@$&4 I MiH@כJ+%bZ' Zg$L=28y6=6 O)q%8~_R]J }$1Ӱ4OL>U4l!WHdY8˾¸qbY>u_>tgUi{_GVTprd:\% x!Su,'=b ZG؂2Z+ds4+awAM_NW9++v.pЕ eiJP~rtL]8Y ]n ]MP:tut!`6inlNW2+}q!J75d/aAU3Ĕ֭.t]䘖dV+ut%s=?1M.w .tYڄ6 e6L`t7ЕWҩ$r-DW82t%p_zG+AҕMC SЕd\^$5=^A ʳ\]p ] \UJë+AtutĔBt%/~&|&] ZNWuJN n!lٻάBW.(Y5U[!] +eޕ%wtbԛAVMa9G8+]}J!`rl \ZCAJP&5HW9iv1@?dd{N==1ћ|7|da|!3 SWHKƈ\peJe hٲ+A\29%k(-CW*t%h=:] JWWHWtSM>:=5[%_AsAxɦYvtcE;HnSL}Qݖw{Nv@5ޤ'*ɬon^>ן~z 1|zՏyO^a?؜~9ĆViw.kMXeW@ J+p^ n ] ڐNW2gi%֡+ƭBWE*d-R^7eԕ=+bLF__#].7mE2ѮDR&tp|`w4&*t%h9:] ʜ{N=?1&ؐ.NW;ܶ mh6`b/zkwy!?Vv\ft%(YG]uЕ{+AJPNz+6/CW*t%h2%tutE6' 9sLѣӕ%Z>݆2[++vޱ]sXnZf hkGNW+f!WWBW6PUuut ҇ 2/.h JݻJJ]%W+A{6GLTUrBtɤeJeԕo ʣp@WS2!&_ 6Bt e>V{@WAK}+-CWUJв?:] 蕮E>8_npЕMth(]}8 -sȖsU\M,bpϓo$%4·~:G}bt޽WKOlYL^ҙ]Ϊ-xkmۻzaYSK:?{m<|B`g%Sg.4l~DI|?evEgrUg~7lN9S^hOó1ػ6$W7< d7FDOgd8d%!ERFe3iVWTUU]ݍKL#qؾ4b_`Lti32Wa|UbWb[Ⱦ%>`Ds99-p'YNg[~(qI8_\q,J +lUbWb[A]%.>/\IB)>i0zW8~ >I AɴwrԽ0.vSma3ar&oWLY"V;GJQ&!0Q8-=bCN_kt9Z#1Un7;<׿}7-ⴿT|g(DT <6R$`;ŨԪhtCa,AɌpci%7  /IQ0p.GYeN1g t_:W|B^LC]?`\O@\o_`0qY2cT7gE? ?Mtv{Y+EGzoFH.ZWC h?oRlzxj/U;ge7v洲 scrT>|G9-yicTdi/ 5+xh6|xwX#M򳐍qbaYD=cf:7gpq~:(7 [ch!fH |rVkSJx->c]dw vK̇kvB.  ,Ӳx8 67F/ \j(a6z) OB,[6 ri?φC:|?]_xXYY59[neE Gt_И/UASg5<+5,5mj,b4?]0{ps&ՉWa] =,{ ti69H;&⚏ &G\l9jM4IXKl0 0(Eb![cI"!b=bDx&`QS;vh\I?]QNyB`(Er",qTc- Qh$2nH| N<Pq@&x92J#"( -J)'RHD#i6O*֭4[tE,ꮁ5> V!6a3h֎Hx$݃q3$/`{@o>P?m (݊` 6O4G!*IA?첷{6)e ~f$}=G@R'Wڄ3:ˆ^FcD2Y+냉KhA #(g\D^W#?mˁ>?+e?'|yʧC>]%[L*3AM"i靔rrԥØ7o5hsg)ƟǠo7cy*p??%gn]Wƪv=zD'zF)mKqp BJ*TP!GӚ,£X$aa8OgH` tP cp{CMc (sOD&h@AUÀ%Tb2DO L᠜nb΂wDàmQ->YGL UEHnWtLPc;Ӽn`K ^d @"DcL9 vFf]\ypVEtJ3 hCuJ O;g44,OÃDho c YZb8ȄxFѤl>m#Jp8hӛkJ)k8a] ~%? x}ƽG1fYRy|V>t=y3GmS>W>l̎No(B}ؖ]5g 99BU1 N>}Gl-NR۩Caq :ӹ'l늏O P]v^v 1wSޚb/*1QEI"9syn$qgi7@"HL9|Gn>stXx5&D43YdSVP4Q(Q[GBX5AQ$)>C즕|gF5ji/"'^q<-Ci!"rh Gq(XJűsbiHky>6FQtӜ]uVO=Ϻ?ZS<|44)d^\gp"gj&$(7_-;̏b80(W:[k^e/~8y"{q6/ `V0ǣA]-x*$IAĪ[nU܌cyܱչ20`4|ߖnNԺqV' `hiiy/A铔%;X|V_=ϲw_&}d>7 EE6FF}Vꍞ/i# ]HKhq8&8dg1W:U&.oa`zx9 #[__ ;Ar xu0_M0"'bA_kuER^e%؉0!;P1ESL9g߄k]XN4B"s Ĭޖ]3,ǑII<&w߇+TI^Oʝ;iQ\jpq'6`$!T*˨w2W…~J?|󿒣_MV߁[MVfRj/(eƠK?oJtA"wG@s|סM{IO-]|ƼFS?b[ct-?ϗg77z.ꘇ}jһD8SĀE[ `zqL@8FōBRb&y$5FҎt셥3yjY˸5vv<\,7K=j0'Z` H>Pm]pkIg?1[4MRK ܄6EkǾ%Vq$\@KeV8e*"St^E(B|DEDM7{,V/.W`˽1[29v@\d"-ArGsQ UQ)7tltl)@ѫ}3/B" ۸քҊؖ- ~F'5y6rx}UӛuZ.=[Ywmq6-cxK>LVdӔ7K-E-T/-Z"VwKfK9jəkFK4Ίh b Jfx g\{pAwk>eP>գ ORڵ% R^ky3쓿{?&&g&+YM{XmG64%פ'+iD<+lׇ[-Szo>U6XxWmٶbMzzM9!iOfu]S`:V\f?.4hdSӣi2@7ɶm{ gTePUFL"dϠRfU9\o`Wk^'cJ*hVp"0Kv!&Ζa^k!{6#%{>}Bʥ8b5mR @.O2Yk+n9HYlNnڶtUJ&r5Q~6~*5 Ii4=^sΊM͊:=\YZ/Z8<7Sz͓ǖN1K<*(Ѹ֗9y9sIǓ]n8iVez=WO|}P:6qtH"ho1;Etv`]FEh I<}%_]iR8%vE 'SK.LbE`(Tod&fdgtn3XhG,<(,\HcWp/ζ7LLtß7f 'g9/24%oŪJl(0CIbpICsb SҨluRTűv*˹wFnlFt6)K.澠vٱ/mG-ص3ĨU |8t98q",W3MAĶM늇l9y ċX`MiEt>[1k+Rb*lTW:a7q6̩9k7k~<]gDt#"nƇ=+qj.(B!.=(ˍUHUE(;IYGWC,47I%eOV Y/Qk2O=D\%b HWCru~Q@ݭʮdmXO!ЪIXIY٩CMB ra>&]K,=X%#TRH H2`u$]GQYY2j"~{B_KEdH6`BQ%.2ee61G!f&Z2kU 'b`W Vk098UHYFpNk-FQ*Ϊo7q6kv3rqvvɫj.~[S7VF؎6šk ߇rAY(lPGA!uQvD{DeY|pQ& uRfg[' x'yFN [1!$ %fPpFqjcSH:k3}#_8bOC@j?9{zk]P,⏁XբV0 (& yşsbAYeg%:)d33 %d ,^qVAv;} wBբin.-=TDL]9ʞЮu}.đ%m0$6FJkFmЁ>i#پy=P^pT@e%m5P>'2&-*__qzTHC)4rےVH%W*du 떓"UHQD羝iݨ].qRxK͵ƶ#^}EATf_큥?F ^?L_op3]_x.~=L`A*=cv0`}uF9$YbmuUz^'g<=E+=h ^OjzSuNiatz]{̥Y@d1f[WZ1%Z\20bQ;hx`gZ/?KZc9F3qߣ>xyЊM]50uar[0[滱V-b~0fzȃ1.CPh*!z?)3,:19f--o41Ķ$փ34.`bv9$_+l pZS!T.B"ط8{(T "0HO„b/X2e".M]8L(>o| p*?q͖$8O7gr(NBm$|۽qqu֑~zR;=O:a`ƁuEc188`:17x_s[a -FLC1R:o`L.fLhP#wP#SoF`v0O[kg3 ]_wݯ<ٍkEu%8fO_/sVzֳoyL7th!8 -KnU77w~UhZ-wuWןm6s~+[ySo|gߥ^s׷to<:'OT'yno# 7<ٻqCkU,MCb3$q5H<:飓Ne&X1E4)E,cUX$ ^Beo' A*5XєRHm0Zb-|{^lq|3gPk܉YX-W[f=}L}33?_Z6bF{uGujпZʒxBA4mpX_bjElƺAP7 a_G"P\m)I+є@"kɨJ̚kb0cXuz9 4OM4ED>&/ʀ ɉQ/*YLZmj &y9}m<6aJA-\1WQR|d}SnK~|v=LJggF>ǿ|ց2΃]}ҲE38d0Xp8 2!/Y{Ĝ"Wl]K\S$|I!gG:bųEc@K3XbcR˔ Lf-ZeHuJNigB֝}nlA[LXK&b{:%Ug*O"&H1z;B<}(Z~iզp~'up,$"}.w0@M2E'?.yh[JjDchj4Yl'[h.3(~ubxYmbPi%"E u{?&?/33Ӵbu^b ډTh9ILIŎGN(<-cTBQn~S>x 4.Ey{?3,LϘƤK z[\ÃܑB wї`&'sLdؤAĎu_Žr̤B8 b.p߯10!NT.,yf gK;]JCMCW'HW0NIS'U•.thO` Jk:ARV.^}t|+Dօ-7NWjeJ[C,]`*B6th#NWR4oO,L׆1hqqnXMP[sj+vha:2W;݂hCWOzJкFt_~}.th:vB5tutŠk6tp-!u+DSw6G{z )ˢty<u=xzkT(\LuK7ʹ_W;I1]j"K͝"SJS&E󜳤<uPsIV,Tee 9׃ipgbxM?<Ƌb'K,i:_Ǐ#wu'N&!ra1Dx%02i9;Yt>9^\5cQw|ݓ -ux_${ٲfC 0-B~v$d(R>g7Q( H)(8ю% g ,߮m?!P"~7M |?KǤ:O;Ǎ4ᣱ&B'-/xk"&DFE+=11=0 =0{|][h iᴞ&FU|C-DZaGT˯^z0Lz11^wϘbrϽx;پ,0<}p=un/hbwHl 7?rf`j]RLM r6ٍuƭ+!eN(2ߩd1R}h4]MKV u`kiH` "(]4v6*"QIVgs@Y@)ڳ2Ǖa *n#xx;wVs"eRfGZ[D8-ot;p8QNG›G8V B×4FtZ5Ōo- ]CUCqFD: ^w;~}iBBֺyAG>q9nBa﯉ovu~4?69'TNg2 ǚ4Ӡ#̤Z0/iL, MiSڔΒB->L\NƄJm?>phtv#t싖[c6ׂ\rʽ=+&`HBH& +<5R,+Q4Fzc>l.*w\zG'f\h,OYz*z%KKj옍Ԙ`3&#T$2謏 } F9p[-IQg#}op6x t4}|5 ܵ~_maWtTn/-S=wL5(bz jUbzuѰSBǿ,ZYG0BPG <- %,'Թ,鞈%6\U(b5.v36}kY$M s]Р ^tN8ֳ@%a#aVLtlj(4LjM]*d$r˹L*&&GmiDg78NшT tBC+|u gfnIqQǣ. g+NY&s.RnEq520q?^4%Dxə?@@ m{=uOD Lb`Z%9m.J0E6Vi4zDr{=ܶlPS/U qeKAUBGp :)F.iix{34дIG2P <2'J)$9ˡ7:pϠ()QĤI^dXӑۑJWGPsE03pC"6$N(e0C|N$)OZDCN=/VPuF뺑>d}M<.\ضeUmN7Vx 9)G"[03O|`ޅ/;hBsZ?\oJSyQ x חD|j3&5½wM " kHB+)m7+߽m@WԵQ$9-+ $V:mrT;XK >OVzjǛ4+?]LqrrٗtH&%&w|`Rz"g6 IN5I0x$2k50織*[ПZw4 ݪ>_7?/^_/',%`v}rsqY-]pAS|D޻xsz)vȥ) 5zU{t Iw {I}% MFl7HW}fyE?OY:U,)ĀxX.oZ1pX^1Hw;<1`$u✫7TlEJw Ns5͔i($, 1^0e>:F ,m>yĎNew8lVP,:9@3%)Z 6;{`;8F;:&8Ȇ;gl+!7gm=2s˗f+/4._Z5LY$fk74gt]/Zبb4{X$dGUDM!X 7$m Af1~Y3INh-y%q09g?Hɒ$Dk Fh1"`D K"Y̩1Z 2eV1 OD1eb"gCh?|4_:8yp6=hE8~~eNY}?_o'A8QR"Eb S6CHȠ?E3uS-2W@BL"x9r 9dBrcG<՞Jb,;A]`{{x'e{ǂ̝x!#͊ཛ'S`]jum6kԻ6Kinf4 ٮ \RrN\>dQutӟ+x>/&lՋ=mdre,֥D0X*i9b?Il60+m$I5G5<ػ^`zDdd-ZRcIRgS.`ZJYYU_~Gxf2q}FbC. *8bZj':K@)Jz6mJ(56S-J֧CTua) Ҫx҅+0٨:E1,swnyL),9 5ER']M0jl$ Ut)؏V;YaJLuRPeDNYg/4擃U#h'B A(`RmqJNG1|e-N úƦ t}y#86ù=35_(&ǁ;x|)~ٖwV.;8(ՔCՃw mP|N*nn=qZL k3a"o#T ĎD]%#KVEtt%u9m#a Ip٨I(N5DA}MkFnGQ}5XO+6m?ƺec/D1{&'MMVMbhz !'?Y4T"Hΰ:Æ 3l̑],ȫPUA:)գ|2jβWD}Uqjv>_uX!3m9)o=ƘZ`7ZČ[eRC>|L9%IM7 0<0Qc.D }&fE߸ (%䠰T^7ʕ ߲Gsb'{7#h۵\ kW(>;l8H M <94!eSs!$8 P.""&0ZQc,>vp]}.n Utqr[vm  mZRB6ߏW=ĸ~!/m}j{fd3yFn֫;f5b¸st9mc<ݝ  nԷ'>񔄜t8?/9g-!L)vڈI\hF~qs'!,;}4W\m=|">O3nWT bEX}LJTebC D%N..՞ŷkiíCw<{^{ !iaM]NӲF1Q;D}qm ^\rzv7e{ƣʎ_OWRtߑ=V͆)r5y |eͨ|Paj92 ˓!Z3P\*Rp0 TY'nddq(kcn 7M6F^TF]-F>eJl9CsrH.3Ʊp#6\ȧ#^:Ѣ>>Tk'.knlv!qN>7|l߶+uC%e~OG9+TlyeVfR:QTAc^R^/gNGiKBj,n @xQ _^Z&5;YnH ҕ~t|eb;1bZ1i]{)?qߩ*Ձqv kpZ Jr*k+ڣՀrՅr h*jaR.ʻeo߰uᴑ<_w wwnb7ɾy dhjqT-y{_ 'GgCs5K{w0t\-hdrj &jw2tތWkz>|X<beFFϖ0b)1pU8Q!)M) ɗǗĽ7xe5J6b;k}QǑiõ>+D^`6U]׀xNe܈j8'CjL% !8a2Xpb)#DϦ7F[;_j.jO&@rt5%kmB]ny%U+UJ cJY,ŻjuKw!VDImbNj!+a@wv=[>=rn{ȟ' CDJi-W!묓jcqY8J¬0(.Vh_sWE-Bq=i/PS.F+nZ`DIZv݊)6,ekg>7ققswO_Cɶ,׷ҍ麢Gzu&\G'>ZtE79D P=1E69hok@t uNnE"bZϐ61TJ[T}=% ]ͬ cX+`(eRtF~\vbUaŌڽ5f<ý9v׿veOVs€GGOG7ɩU' PۢJ|5g+ȬY3 m14c(Kdii}8TTL@*|ͥbg6sbb jw}Am?`wJ<# !\`)F6hCVGE5C=jBW<9aĊζ.*{SL^ciW*Q]OunӹFyYƽA^XD 1L8!.O)d NQ+_[ذ;aJr%m\XrSSP|:sK@4ME^5 .<Sg\.N8ncTP TIƠYgjM6gac1ƃ!uN E &\| \<&qxvGqMy7Ulⓛ0u=~|Iǘ_"Oo|X>Q7)VyڔSBJ." [*&Q6љ$4J%if %2uFe &PՑpMVH+UD336{A. p\|G1z(+M;ĻsTƞе3xvz9^g X yk&Mr3&O־zVԷNINU \h/XX1U1ԨFAZP`ޢ(UͪJܭ7P`X֑cV>5 sܮ,,vKjYh:lqA:bT*?0Z6#)q.>]^]aBg>GSVQQWV2aߒsV{P(ڙ88zKV51X* 6'dbsDtf*r:$AF&s;3\DJ21' i["s)TWWd}8w xUmٙ== W|>e?{WƑ8q!vЧ5E2ɲUoEۓ i5U]:* 9JEٛhUU/m?|E5/%\/_NVwEՍqP{/Ub:|9nW08<|D_U< ݎ^#CS5u\9q~@-Gm/]э CG+g~Q2@Q~~nǓab-9:-3`!ѓ:Q1yV!b (PSv7 E>B'TE?OE)^սLL`oݫegk{>ftuT"D!P=Uyϫ΍iMdl }7) }0PalY!i #dnj>YC_ GTa'Bр|#[UuoKԫ.bהgLb|: 'ǝ,8(37HfZk"+?}|ۯY_ߏE9CEu~uYSM&bG㝡ORH}},'e:ث^6c*mF`EO*a>aE 2~zL }]ocLhfQG RPcA:P1P< 65Fz! 2Fo ƃjǹ2D<EAkHh:4lB+yѿCN#&*jHBȢ G]Kymؔs#tfbԛPDHPyL4:GġVdX 6@U!nd&AxW<ƿs#. 'KWR1TO AiTs>Isk4dR>%EvcCTL>^]cSJ*]ck#ky\c*g?SX__X<Vyam9^#6~:^UnS*Ĕ&P4CiY  1i#Ti2]lJ7bQ/x"~oepqwz0~6>ǛAyO~X9'Qw pQEQskLJ ٤C҉yJnyЇO\)A?SJRL___Eߦ]߁ڏM9+p +QLZmn4|6g:s%3B'e9 [Eo7 {W*}W(|e6P.%C vNӺr ]BT6pApvفza3%#C4œ, ViQr9A(k<ݗGa<<#<*ͬېu@9&)\D%TFL#$9 j0Sofׄs'NDFsON+CAP"^QYH2C"%Gb"NR5ˉHQAДJ=>Q?bkSK;Am._7xYsUaLqs`ڋ46:?F B5bvw wFClzIqK\)(:RLI)K#&?$RVQIT9YP(OWHˋBVI6Z% oba eyy:UiIqb<\oْ(3̅n|vPC' =vc5b$ܦ#X>Ūaa@fX> #g`,d٘ʼnգ2~"Wj 3*_v $CbͮczD6ܲN%QD ;~zr~RzM/ЩFI~}NuNE{GᅝqwN8YDJ}No92_>{wޜo}8L޿AYx2\px:5A1;C{ Ǜ 5l1&W=-U%2C՝]a@,W7M߯j'<1o#|oi"j%X/[ m38 ]]7We6ɵSSMf ']4?2[ `[ϭQv NlzYw-Kl9}Nϳƛ;_Sܢ繖!x^zlVsg9P˼5n)iN?ixymzfzk_. RZtsĿOp(nxs\o޿i5 2W2)նDAJj}*FYrvnKD&n=1 GWB0Q%ݤ|{jH]O*:E DK VH vGmj:[oF7Pizǹ܊{+G(xHm0.iĴxBTxՃY !^&iT R9A&/m} y+Gbx[)fUu`(CpL{ ܊`Gj@OE)] n._4m]+@[8/)NHP>)BYJh%} "e#C Jk&DJ92Mԗǹ&ݴ״ܗA ,;Ne (KaR$21SƪT֊j˩@ I-<ymhSo?ۭ:L ٲ!2J!8sRF[3jTd-wm;%eCe]coW쾑o3![֙gZ344n j=⬛9JaF8g*R K@VjG!>=! MS׽jqԷ[_/j~]-jQFZu t?Wm iiI Ny2SXU.K&E)a#bZsXĕ:VwXXjW5-ԮkZմ^c4D‰"HI<7(d JLXpBo)}O}A W.i )sE5B+.aJX,-y&cjWL$`gϋhHAm 2NJ`=U&,PT |ZKS~|'m4>uUU*Ǧii}_MEh6 :F;AZr\+KQ1DŽ)&ep['WN{NAOlJuUaTaؖ* ;"Sks;Jb P`TIKl(A{U ٻ6r$Wٙv?vvo 3; YE[Ye{|Wlq[vD`TͮzXE ߷ZCeq?.*O=obҀCd,dpCRΌ08'd:dǒLE!X.sz!-c<uvg@vy Y%%xf Q/RQkq9YbI#£ B{`@xHѩhB0&E :).]Ӟkɲ*lXQ 9%O8QńB& i*苛@03O`=гG/ ydDkeƒly! ^ry+It2 S"7d0mbJhXKpRQvH[aZCQ ꃔ-Q\?@jX94^4zݨR.xxNJ%"d&Kּ9/Wv>$ryQgmJhab1z˴2ZTs;p\*"ʜhAss)a)w(t z{~>=0L%^{yP:I1~GYaB9c2MFg<=-~=wFs__p~EvG9q~=YuJ(.4%!]jV)4lIJRX" M> K<gc'Ȝ2ZAU3B M?7IG4lAi2۬BW*~υ"[EM^МT5vN ˢPթz~yBCϪ\7MS?dz9QY9]^ת7Yjƈg/'=joߡE?<^=aϦQZYVQǮW9%eKIE4>}U8 km{z#Rt,$ZٯBX*yᦿ qp4Ys; ؍-}b "fQԂ$\S1,N{WpGVNƅd3vp3i49Y:8ОC54f\FmkM[Ҙ5mL&x/T'_-QIP\j =HOP @H:vR)t;BQqD/?T̪|rMaU; ˮ/}K{(nhN\^ vX: l! `Eqr?^^>}sޞs,ς&:9*d1ޤ[$eQhIUN @HJ[VoddJ턶D.,5%IŒR"JYNXOgΆ_w箠b$YTdIAĨ7 H(J&lJҿ\FRҬt¡tNxn2cf&! ں!qn1-.%FtB!d ܆ZڭQn-W-c;%c+1 ,%[e(9Z$S)&˱D2۔A퐌0%UjxY{=37=a}V9{`d4ƴ=Za[/ŽW;p0YJcq (E g'~?D~mK'+K7802 2|I^e<1,5ܤVT@ =mX&Φ_D!]>+|Hznb NeVG8&M\zU\( - #؇ޅK~Gg&OU[Ko'o&;!hc\Z6)gw>&uPrR`k**1_G}EqmQ".&holkk^>8D$!Bj4t2^P$RYuj:ڙ\Cɶ>eϮΦ>(Y$=[}}#hKrmR'Yv?r&Q  \8.dm"\$ˍ6/,\~4lY9l1M>L.df ZL 542682bcg#*6?/q ww\\5i>n'kPQ3;XB=BEg \9%ͳ2 &HLa[:MC1&cRؔ&Q`ɶcg<*or̒u#`Yq:sWP{06{kr< wL[Dp[2ТL@ŜT6 31eaì ):]~5i\$Z(Y )X>!k`j$}j`<l85>gwmv{D|wTq $ْ@[ýAKǙɥVKf>Z@\ASOgcD.8!&РVK,i TF ҫͲJ.u%p}ȝc*gf12BO쥳Z2Br!Q" Qj2 ѐq1pqWp0<#@ؚ7]9v\YƩ;7 r''~OC:!t\~]Byq:l8!^ E4A]vm6*l7Rz _OX @E4RfM^sY3ʬC ġ>Gvp6Z`q<%s nWx@CSpg2bv,V $T9Pz W̦yv}ﳕ8E6Ega{YXyWLN,˟1MgP@W~DܲA³ѳRir14g];G tG) ?n)?|h-sqGQr svdIY饊hAh]Ho >5v&۳v]h ЅCt> ]4m|t|=-/MP=--6,WݥGx?\[v|Q˻UW&MAW/gZ{8_!)fF#dm" l/ܞP&)օUusTCb81Ձ{iu;{󫋭ۢ^r[Mtwv'h4lZ(+^ԡ١M8Uqʤ[~@uPT/HvV^17e~0jMw+7g1ab;ُh>cKʄN:keCo&KHqZhv8]HɴsL??t1I3v}oY{ў/VHpHOֿm~liX㵹&9Փ,BD]=6^䘬VMlevJYm ֩zC>iAJTLEkNէ*U+YnUEu<#Ҧ*[ZQH+v=h-*7snpZTm1j\ή/nmYi%m7Uu`:I moKR#$\KSmL΢3RN›1-.nt.Gq)b/9%|=„;jO;"J=B2&QaQ4 ge|uJYd|HEb5Jڋ:fxyF9Yr6x}1>5 zU=mj*uoD $YڢDwAG:V} w:U29$Vh4v$ԳŢKSuh5ST:8S׾"znJ3HT'.YkV it(IPkCvYvӌT"yˌOh<> "[mn%Wl; 4shq9ݦGDnDM2U _.JUEʳǚ[]J2qF94VX4α!c\vl t[ bfsP*Ҷ4p`G4$: MjA\"X! L dHS᷒2UW@,TB24V^X}xeY r3xWh%Wk1h a̷6 9 ! Ex&P4vC23%!2xk#N>KKf`&bD`0ciƗ)I1 Lu>% ic- `I. pqDVZ hަV J8 ts$`';fgغ렱ZëQYQ@IAPDV7gd0[Q[jkIk $DAX,Z v/UT$CwXZ I@<38onp}r%X UD1&"9i`>b !Rs{r;'iƮ]-^FޅH}6=DD-$:`ɖƛC G\MJ)h+F@+Xe ,Äc*(OpHv9Ok=ȗXP p΀4D. i ^ Y0ⴱƣ4@^BB$>yT_V 5ȤuՅ*vdl.f,XN5GXd~^j~gE3#md-ŀb$Sƪ`|`N=\=Ŭ<$Jx—9t4IgmDT2[{(%g6ܖS5Y$a @AKI*ZB(!3ڄ`btJ߳<3PAȚ\\"(άA`&#E4P]  V;#2Ϊ ת ¬0,,H1dlDH! c#(jS,.6'tXI#,Iƒ5$BYIe6Vӥ*^ZX@hf %l@M}w:6E-)K M ڪmryXv{yk99eڏ8MZ/6:b&jpL#[*6ܲJ''40z63 6Ilf(mT͵%'ՓFCmn0& |zPӭa#·fĞT2:%<%2`rh[S|'Ѝڪ-H׳vJdEw AP`AH Ш3ES'2(X 7o#2l)b*'+ IISL1. xrSpM9,RxyE ap`R p#X("#:ztϬcP\G:CUE :csjo6iڽfa@ZpYPAlRԞK$^˺ytı hsO:i:- v!.ě"@[vgZ}pk 6VP8kU8@6'=ʤh^H0#"I ljrV9ކ`ݬ>jE#VnDt0r(t#jI:`&0ts'T\- &!K#B'5b ͮhx蝩yDLǍ o`j?ozB/|7_GAc p0! |floLjA =z(j a}VD4&yu@_UD+><%%N=!%(`V:JQ b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V}J @!'(ا4d@ָ^ +g%GRR@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X *OH Dd@0KT@d@dլ@:@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X JX ̢SR`矎Q|J 2cTyX J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@@,;{KZjJ׷~rj_-/.[s.dh58z׫sQ&qղ~r?mf:ˣtlo\ώwڥ"=_Aeur}h:(m4Mfxu è@\r+mX,pϳūWkSKJ7Y;Փl$ώcVu\C A!$GZ(UR{vu^-z*b/]G4v]xalC٘^x{oWr<*D!9Lhb;xMrhQD.BʷV&p+?,͙d=$$31qn(7 NFݶ, @J-sjQEbZ,ŊsLN'xSW{rs'IPg!;SL(~p@dUm5\o7x>Shy*,dzآAX/ Goa)$֢9Vr4np{nU*QHJXhVevsT8$o篶7`7ݶU۳orϯ2kG՟UAU_qmk[`C@&R.sot딵ReAjSʚ'(%d{Y~p$a| |w5mI7Zjћ勣/?3~{?RIF/~g>v>T};~HΦwda*11li3$;O^^_|$(ޞvW1mdpa0-!'J_ʹWunߜԎ_.tջ78 >W|.K 1=*/zK՝bgPoksף7d{zX 4&NIG"taUE)_ W*餓݉G~xg^ics]6{-_oKq`2^ w hD߽~oYZڣ+zQZ*}jw/+g{ <}7&Hn쾵ڒ"ǻ H|ѤKZh2[} $T8wg9Ýwλoy`߽,' 4yroq0)|W¶(wǷۇMտ<;ߚ\Dgmf՞܁6նFPOzE CkhN\l\`5?{WƑ`qF/!M6CE>l_%ɐeU _%qHjI=6lIVwLWTUW!{_)bI:dS EmL5jA5Fe!Onu]`~1Ԍh!;Kͥy?gyP=lfshVT3%>dVYkhoTQ a-ŦBؘ62{/Wւp|^]*XeTc1ߙ[&o E̮Liڔ)<̺@&1qevm jDt j:,`iGg+3[Nje!dފJȫaï[pa/dt;T {Ú4~rvmKOH6nO<4v''^0psAM1e"UJJ4 iDL0$cY =3PmrQ3unGL$ʘ"hSXcvB1Ek[U;qG c L!b,1;}"+oYEQՇY*7K@T`f;4蚘HP,1GQ1KLnLևŦ]6~s`4bFԅ54b7xk j" g)DN%fZQkQrCʆk"y.KT(Mgڃ^3:$[gU^b7x`>js72YƜ nsh S0%g&.@&BŧЋCчv?}*ld}_ӄY3m<U#_^Eެ[!aMAp] E? ,S>х Ў_2'>xoB0HᰉɤvkffN\ : 7y0b ^j. eT`-1.,D^$13YX#,YJB:N'}Q\]"vA\]l:;A XXs&9ŔaYT1QYTbU>^6&:TG7 _Uݯ~|3WۉIKoVj|Ұ}?wU+=}stogLgc ռ pp 9hu/z}?;q `L^!|Y[wن&d-^oU5k&2ʛ|VBP6)=mL/Uk9 q%I=cĆO 'j j>7ܭof2ٺ~HW FTGpjdrTQccv%?![< 7BAH{Z[=Fy7[ףd jѣ])c֜nX*15v8!L8]xq+Ú|!dTLY3۵ -,^8k_ g yI[.?<⬲Lu},8YMڔWv9-]U[wCAz̸Z,Jt&I!(O_~F߿V9Ъדfrz|0\y492R>`Ѱ66wH|wgBFbn[џdXOhXѯ(@eiX]f;ۇ5v,'!u pU!-`Q(*F J%Bs,n$ 4LjM]J1Z{c`%W. ֢u-CsMgG]`L+8`r&a{i-V:U:9dcևk.nKeF8  m̍zZ’@MvRІz[^eװ~}.Ra(AMA+sdP@4%gNzhT\JLks:kCX!5ibɁ $%J`0Oǩh"M%⥅uL",g[s>KAUrO*l3[, zhfPCQ{/pL"~F,%%45ɋ  襹"xZpC"6asxQʸa:ZDӤ`Qi0^Kl"}҂w@alTD٨hP P_dlIU.].аٕ9?=%'ś\,T+zm;tBdUM"Y>e3sXx- ZkFtKۥ=д_3(H=!:i#1FrCN,˂KN,$;_)) 5LLJffڔ ($Fr yg8g02,6-ó_(,^k5&zm{}o0 2bv.kh?7w?͖4e -evv}QK\=ܼ0m{}{eTox:,q2U{wX۟7]9^yN:6/R an7!n.87oj= ݞ4&GE^  <ԞWr B)|IHtȰs2<&-vTP Q;xbE򔥧+!\ZZیר1Hg 'dSV2_5#KMgT/5m'BZ{1=+7Ո\q4Em s: 0.VALW_Jixy_R쏧 c/1ʎeft"bF)BTjU;ze7]XF'g0n8 A cjPBU.%eKu),Ve`$PO14Ŧh/#" qgڏYÙπ3aY/]A[>!Cq?&#j-lC@c4TtTudԂPÆFkj6 Dg+h@fZY @rs?i^IgIDΤ8#4`m8:]V9|kjcۣqY <ĽcxxLQY ,'J*)H̳9r@D)r:SGVZd@=,icQP@3LI@F<՞Jb,;A]("Mg/P$U_Q:V0%VlF[|!-:5f֢۱ .gin/]%A0< 4#dOnPg+'f@.1F˔ 6% L:DGS.TЮ'RpK !!y06rXj;wtǹYkØBp`4kV8 1md,**` f9f9򂍬+dNrᖥc5I-% z#VCS}cƘ Q;f9yb7y՟596  Lk1G: @H˃!tXB[]Y_x?Rjsfȵ1OW6&z4t$AR*ݻFFZ>U ?G#-X~hs]/i{a3{~'.m{ہ8r{v6MN"Gbgæ:Œ_n a+ Kr_+Lw*CT,X0+T)峋d?}"͇`< hyna<̷(1bBQ.,t0MTju{+ek)&D*i깯y[j:6Txi{7&i0{LbI7B 0A*1j 3ꄷ͒䙳s\şݦv&0* |xGQƼs4)'s 9X1^1~@ E/kICߚkо 3Dbo2ѫ` "Ph靔rrQe9Fb!%C$Gg'yq(gR\_~y< Ld{HPa)Pc> A1:)ffD8DHC'O06.蠘p0Aט{$2Ex |M+5C1IqD~紾Tz_ɻFfjՏ(Ϫ:xn;,x!zdugi3:0ʃ& h*A0 6$Pt'x9bA&N O|?wKqkFMxƸԈ W2vsqa3A^XI]ޮ?B#fZ٦GCRv ^JQۖ]tA/!%:/\(EA(ǜQ< { 7 )MH BKc_IA6|IDf^- A>V;{ϣnd-✣Mo.8] l܂֯7]DAxyҮBxvbDӱq!;~h}V=x7A0Yy*mPM/Mb8&!_ţ%77RJ_+. Zh Oǟ/Nğ}~OW(CApLSɱI-[zՅeר&%ߓ:m@]a}:ͦs˨s7khd=ي^4嚓Xe VGQ=):prQ{b6J>tyHUj;xr #TClˈ*jyԡp6ƓޗJtxHnmo8,YniOb7ۼtnMK6QӂD͖ܪÂĽy]_7 v}N{d7<"XSVP& ":j(sX5AjCAunVnSG6^^hK-}ȫ)瘔๤jQ* ̭01Ђp|*M'Zz]ȧav5>B m^ng> ,-G/^f %|BDp3e%I &cɸ[рcgm\E>ג)E 0 %/zpf<ߒ-L!ՐYCbϺk۔3+Qsv*ӨTR(NB.[B h5@00WIϹ|2MOtyӻE{d%eR/Gc.Bȋ*[1m_zhD=FΨ%]"Xp ,UNo 3~ wԏn+5~TNh/>_g·(FiV|ɿK7E |>Hx nXOsHD(Oo* QNs0(Yk=J0zp8xGڿ]iGU}ު1[R퍓{%2THŴ%3a- L{D4Wnz+5c*N_p75EsJeA xL 0h0x51lFtGg#Ar㈘rKw9PI^r׿r:ku^{g:R"12Raf(ތ;$(r F@}w~, " ӈ@c.9lVԛr>1i Fΰh%OR=Xsӱ2|鱎F"襥A !z"hsۏVIG-Qc$ Nb,An#H)o S(. R:NN0&TlԆl1 y Fmyt.,<.M$)eU﫜I>5#rf:]oLB 0Lg5''Wj4g}s|0һ.a@]4<اl̪͘`@r#ɅbNts2ΫgJ`pVtb [K BX;/n-E~wJgXG&%l pM!HU}=WH`|QN8`hk IM v7|1I Ί˄89NݚS:+'w>X~-|_xm \`0{R\GYq7  `j{b|sOnH}7eƍݬ2{؈> G,S 7. zm$z+\/t|~>ߍ&j +flTm"*В# 5 K짫? Ka7Ȅ[z(JLAb$VPa5;X?] v96.4<0smC R ƙA̻]`ubJY5+ 7:A3PXr8ytRSt*/gH`gI`U=`O<ZU'S+DǬZ խ [,!$ ÝW<ͤV)ܜh<+hV0t' wFMRvy%ou1xC@u^n2fjܹd-m@)h ݰ7062]졊ބA^Lxe[q9iz8 P?HU!쫖[g?mZ慺t8t+ Ɛ',g$ber\D!;Hc5Ed5xArk0($"5X@蕦!rXQqF){ +7‭$VA-m=3{9AZ6[NR.LVӳxꓭ8)m [M,SSAMKJL>*kmH u|#~IK 9F>F?%FԐrpUI8% %J;5UK N-#틎9DD!^&eu)DJBT)DAo!o顀'ȗ#l`Y@qxw2Z'HB@C )>QڻUw/VI 'IMvzJ*#l3?E*'#WAqK*DU7`1DRp !;1 @'eN*Jr s)%bYȜs g̘`<B&# 8jS2&HT ƹDlhzt0h,9{fN|{]A!,d8AH-yܦ`Q} QHuNjB 4$id,S(Irc#jO8KNP";L|).0s%L'V>^jɮj lvQi-23줷讖*2.UTU K}RpK(1I]g% lWԿe]г eϨLHV9:#c(CP7yDܞ͟OX6~F{(F,cq} Uqy?B,q\w"7r]^RU"}Sg;d$6/-y ,^w햠E=Opi]I)u+!+A*c5  \_eJhH9pVTfeۥrݫGnzy(GYd׃5~o t-o=z%ؼ~[ 'ҍp]`ZQ%O3#"γ$߫@=z/UBtM b5-L0Q!^@8iT2YZJGՉIN\*qк.fl։5Rd[ָާoaHXl{JG[վjW,agEX#S(H+tRicx,}yvqJŚs~ON۩S)rhR7uNK۫x3ާ[.8Z=ʡΗTG<2bV:.yR𠢷o$omJmWaTaĎ* B5/6׼'JJRK9j1IX؟~!r]/kۣ}[QWy5? UԔ q)p3ŔW0cZy!ni1 GH`)P~{~n{6/WקW/_hK"((16gsܹx\H2VBVa/!1F/zr0`W)|d&-M";ňwx\IxϠZg~W5"pCInsomQH:[ ~a,v)|&TG0 |(MDeUD*`(gO7g<8i=yorfNZKӄKˣs*lg !Rd{AR/yʵ0t?%? .Z!zDQ9w 4Fze! +TwT>V$P$E5ﹾ;?( Bt2zhPE]kvCa^G #ԅ Wo LWsiYϞ7B5,ZEmYy>tmn;1zF&n*lyB?ѴʒʼnFb( B,#gIvIK ̋h '/)dޟPGL_9,g) \\-1Q2 ULFkvDe 餬rr0ڻ'Uk_ rz~'w ^Vς-wI}+EHSl-5-Kn1pQȕ* T (\cQp9X-α&8|X_G?-7X "2bhy.N)9U8NC(gh%I6"#\4XӖk녥!R"1DҀJhhPt$>i>F뵑 6 V' :&[<#;GMoķ|-y1j]}ZP8>ݡ{QVů?(Uz2^ JԷ׿fq[ߋ8.sGo>uͼ]q<|P,J9N5SE9?%{EabSx=]n:^;~߸s\f~ow >S?˧@Q^}bkXM (o<,-F9N'3Yr13ɸ?X|W QLs&~rU,⍯g脓'|#pVgp}coq|]5buU O"V<hxpLP\ۛ?S?7/. }‚alS!pFtk>oԹ{L!Aa'BȄcWpG7j:0O&-8/EQ7B( xq\q;xd6 5Xu[)"fE|+2F@ (*n0 nK6fgzm_N\Po]`#[[,Bb,J8"0C ?ժ?EC5܋w=#?{\Zţ4RBU rww|}V/PXRˋt:P O(QG8t#%O?xZΏ_/̳ۏ7?b5C>w$/0̡*$$yVEF) ^R Dm^ȥaRk(c$r s * 3MEmZEǸ.}0uF 9E)hI@#^ZHAP R]pJ/Q(^DvZHھX# @j jIxx 3:9,蒌T;o tX|2_5Q7/Wc'eđtuBc=eDf' d!%.t=MQ1E(zsoX?XCso1lbMcLcb1P[鷕2x0I"$4DR̕Xd)'_|"pŊ$)( x#w9#eaY"=K4r8νf)dc;i7`2q:zyte\ ;'vGһ/Tyle^<:M23+*ǩph:-( 4g_C-wüsvH lfdjapUReWo$MA#Xh6sv.OUs`Iwp/Y5Z(^>=.Y?vYHߟm28Cǖa˕ &=%%3<˶v{JPO3&V{tq-ܨI0#:HNkԆ0gH<*N )e'Y0F.I*72hR,:pЈ51gmbLPh8B@IFTSXwaM1gKǾ #CMoAjHȩLZ+jm)Jn(Qp^sYD;k4Z|Ʌ)8Iiʣ,e#@ 4;MEnXwZ.\j8W 8&_g1)ma\4.v9>x XƜ nSh SƘ3JK$P :\<. cC¶i{VoI?ME"~|G]8;Hr~j4/%r\k&ӓH%v+U^3C'.q 0H*- XK0H6XiDq4f}`F2f`) 1d<8 fS aƣR 2 H$j$L%ʬbN0p 84]];- ޅzD|$(PR ߡv ^ighes[\;YSĵ 9E0Y>-n?|FF?X\oB͋gwLBSIX?ڄ .80YSW YBQM׃qi.aZ}G FpD#E}00EnOub\) @(N9=M>1\bD:%b[]UM~}˽{1K=yjj\x8i{^ss{?M[z~wo Df8/_ *Wn?F-iJ X%8RT$X&|"% |i+˽&Rq}%ЦW\ neK>Ɇ&gcvNMAl\RHa0)vr9m@ `L"67*86V EH,Ln˙ӥ `05^5gW$wHCk5݆fRq*%5"JqEWD18x wH=/z aү?,,Sk8n2^ϨIJR!{-++,,q_w.Zr}|)5&8(75gqyoYͼSx˓W?|2Ռ!ڮs4 Nag^̭5Tޟ{7D>p~378>m4\j?ĶCLd[\`w{lx1S:]hx9#׾q?Wʎ޶A`^i[s|(e?WgƛF'7!fpcԬn-YZ;ocmT۞4G]2/-0UČi;-$@ 0Kj2ض(ֿAj3fxmqEm92FfI̋Խ/=,>Y^yj=[K+A{*&WNƢR.fK vK&'Ӵުk}.yu׽ ԼUpu C^/|'5@orr6͙_s*n2d,,Re܁N+"#V+a'#'i.ʜ\<wN+|[#d[$wGr*N%;c7 (Y'hi(dxnkt>=rrB:,]S$gϹiuPd8~|B,fURVW:Lt,DLt}Ws.=w\ceg_V(ft?ZSa=NF+IHHt!%'*Q[&&. ZS@# '76?x%-VN 4= \+{n.8 #[{vy{h”WƳ[Urx8oP8AQF>-#@p҄c„S)!U鍡j A<8B(8 !Jy(M\:웉ɨ R;*HUV;e9&e sjB#!:J&k *1 :P5ǩMBY!Kv- ZDrVf#ʘZ hi+ɝ୤ɘA8ZW )x%l<=¥d¥dE{qOpm{!<銞l(z"nEOvqI""K9pr^/H8I4.6٠1̔"TJ8"Jfv%3Z2">IΔt uSg BiY;ەxau >K̪{ʜcɦLztӖWvÛՍ*#vt;Y"xK] BQO]5JZTRM.V]Y4*"nE)OGSŭͷޗc6+fڐsH?-?ok~cD%)<-?>9(ˣ \EM9 S^4`u Z+/]2$4:DCzSz\?}َ<^OmۍIxӑ|#\cŕ36ǝ G L!U#)$vpDȲ<:`x`` u,sJ〟 2IKedA"N1"u޿"޿@o8 jf|]Uj;673f^f85OHRDaE |(MD-:1 Yǀc;)øs|LEʡk"eb,^ؠPAˀ)yD>ɋӝyrs<7Ѹ7,Wq\u/GWO)r;pZ]4_xr\:"+`g2"#:'cCjV炥Cζx{`W%_Xe? [<xs9Zϯ?]q<sXkH:#w' N7\}t8h%8 x~79ivkDcݸG(Uֵܳuޢun9=KT A0?G~v9s%QR-:kvy]CvUgfYnHZ;k(G5߯n!3b>?{ɍ`mZc%'9.|>yaXьVS4#K#4ZX{lvUz4?։ުfߪ%ⶪE#vlQjѪG<*z*ךq3~4kWT*>J@=O>vs#XLTwz:S?>/!q񻕾M;htoDO˳On_ Bc2.MiݞTtN܅)Gkz4[ˆ݆6ߓ, ܥ8{+HpTW^,w)>OJe߀9"Fe-└P[tйdN..E7zU^.]^LZS L !YCE\x댚[O#_(BZhU'TN|r6%%d8f6;4)n ~ h2qA:ydD艿xYBJЧޣa%F358L"hn;.@QO'i6! :f0*4l.2^4FGMm䱝ޟWɻpt=\JߗJݓmI'lIz?҉ٕF['zP݂[b:-F'QJ{HL{JZ tKpm!0eM59*}DX>0t= :`RAE=HR׋]Ya`p0dZ3Eº6 h\EC}RbAEV9IA D2&2P}'A,OKvˍOV?1KIiWTbYF2a 1XK҉@R\VLoɟ >Իy{j@i#Mc| AGU>gW:}4 !ٯgXfNɐ*sd$-Wj-:}5j@-Rs/}r+t{KZҚU)vhRe~R؅ǝoڽ:wr-Ԑc@*󥐘(#K'K.0˔ j%A18!-+Wۜمe0/Ca4g}Qɯ6I8qFеg\|d!<.ƪZtjQ틱"7pԗ=c2E ,')<^"Q2g3K0&띩madt9pȀM4[F2y9?Y(ƱLWD? &ǔYCUl0<]M:jjvcyǴuL%2"i7;9l ٠Yت0kcCI@2XtɒĢIh8RE"ՉTe{X;cx0|jmeh{[u#ʌY(Ζ=Z:Lv7ȝR RqGx򬘑&ZHJxOf@lrX7oQ/ܝ!Ys4s/4~vCXŪZ ؆˓5[J9+ L. 0Rf& Ȭ$43ID`IPԭ=o\pCv$Up\dG%BG#)UJl8xre^qw}_3\)& s5 1K圬,Um _qV`I `JfZŔm}N-L4jf?e>(uȼ* ~f녟;[a&!Cs<#R"{1%7N19 b,{ 'Җ> ĹyD2ktyލK}3okwwWz]X-&t>ƛcմS~ǧDWGH"6tн\{㪆gX'n xN'}ݝ>nM壸W,2#LPk8&Qo0뎚]DP=gWfM^N uX)y{K)LGˇA)%naB@@hmM˖imN8Av»Ok$Jwd d\H?:gGǺ? }.樻$^uv3Id r̝̚Kv\3n;7$f1RzJy<"L<" hzN7%/q״qҒ-jY:Oo:;ٽA)&HHx&L1*`?O?B4맦;=&Z;KM]ODJύVoJnܝQâ_}h6, k)ByjB&+;ի$Q +| +g/n.ӛ7Wb1☧}MoA8M0N MKP9^Uڪϙ/ m-)q!Je) GRd01ZOX3S;[n!<f 򾢿AM1F&w6)b>r{6D򘤇"q-AI+pS*xQn>'tU5E:LL,h rC&-]C #:l!{0dܖ=č tXYƂv7/J1YF+.1 y:Jb6LBd1g]1%(3u ndWnB[eklH5B[c3V0h᥼_zrPac,eMCqJA Eh gҧ5.O)t~AJg]fQ4,ˉiHTR;Jm!)f[DZpnU&"x+FxZI0:p \Qvj &^Vɠs|Y[D; ^ie8o =S'>NVyu$s[^'n*2kmtVq83W*;Eo_a#ӧl,K6Q& 0Y{KJH,2eF&჌Ret^RIy\9ܕTOkT%o2!59T0^ ϒ\Eˉ] /g_sϳ\jz9M`b*WUqRxźz(`gV^00;bzBxU8/~o/Y)E9kٺEx-DY P <dyXn1ky˴iԳ~!UǛd)3$1DKx0s*:A`"?kYA15bjnw=~|0:~,z>]KCpEy>::Lqp| C@l/YYg@K4'G>KQXZ樬VE#C?HhUFTx=V'!8Ӱ Vm?*,-KY{ZqG A(CSW#-12ZKA"Ɍ1J漤(A8ԜR%P99+R]E%ǎN%Q& cb>I޵F$ٿRJ]t>"_Hh@ςf4 %l.wì{e[UؗX uҙ#`8Ri$uuTTdswϱc0?&[(< )nLk9^S7u1'kS-R:#<5uv\6Y9I+\I<-1wBj%j-u1wOoEj9$IO7. z!x2R+ғ++ɘuooS6$Vfۊ0|=c5VMkO4mi^/~@Nvz](k)eQ?$$}֍smblժE bŅHNS@V&թRY/.n[\c$Oo*j :f/0aUVZ.Y6)&o쌲sQg녣ĭ^/\eyPfu 9D+ٮWƼY5iM%j 3MiTߟĿuDC(^:hZvA\ q$lI֚ i̖}^,(DBFt )7oK0I:P42NM_sHI. >)>E.*ٷ^byg?/wq+/N +#8Ѹ{S鏳ƀ?e55Y)eAcZcp`^-pU"Eo""?mV !``U4ت*fiEZH)Ũk)/|2ZIJ)ITjgR :Y&΁+ýo@CE4q:iRQ} =v0s~dbW?BkC/>\,ɔݻpfT‡W=?[tu#U`!wv7@%Yd jx1_nJOQ?wWu^t]fϩYڭͮBtiWKռ)ruq@[%w?:<\{>bvؖz-+Sk?_^]n$f\M/3?%ތCDscrsu 7W#7ق.-"4O9[Ywde?t!MNn==X^f)L⯻!Z1ƒ%LqSP-xޥcN>` TQgg@#n|?,$z՝_Mv_( ZX-E[y(=h׹UB/:;s07-ՍI?E{ӗ/&=o_\pZTN0?+pk'?S6UՔ( M*EsAǡyAM(N|Q>$ү3MkrZ|i٦Qi1U6i3_ɒ MLiJ1iM-J<1@mre Q#ay~^xp :ƒ cRW }~~v_ADr߸5ެQVe?Lsak{+2Qt7_8z;Th tp!dHXlO 2X A$Ϫ,#ͱ6g$GY0ڢEN2{g >c% Mvc5H ۉ2VM`rsa_!GZ)|1 -8$mS1:(,擴RY{E/3tHJ4ap1(liWB37F \rV$l, @(`TJ!Uy=V \.%Ɇa*I%CLe@l x1%[?l$*Yy9mY]7nT3@)b`&[Xhɠ/*OZF%´14$,&bYGZl b# Jʞ˺ APAaqJڬE4,$xsE + ѻZ(Z+!UC¢^8ݘ6 ~+YLBWJNJH!96&T8 "A\˜ŽqWFa^u H+h#1wlՀ|ͩ ~3@Ơ)Jt0b(VH(,`@&eHFD qToMI98<@Aa1gG j>TK8*T8I0{T` |h`XX,!2Ф~ΖJ-ܛϠPg2BQ"9%8_38`/fhT͒ppV@$T8()%RؕM ?躌ְj{TP1b#8j LcSܪD:-d ՘FDYl $DϹ1{UP:o3X+꼵 F x΄=fPҮ`+fc2Dfh$%|B &( 9L:Fх# PdUurraV308pE `A-)J H VWe2'E`8҅aaaSA^|'H.:/e̊G̀ 8v VFESQ_ ~=犊<,9xY#KV}{td ./}4L"S,K0+Qw@pvpC*r]p_.$0#TK^H) e@;y(%G&b%scY x %@@!Mv% K |0bF F,n0ZZ"΃pzPv"(Ï& Ʈݰ6 3)ȂJi8bzcUjQr`ʏ üO(@AV ;K$@DES6@k.OCX+jꬌip`k3ˑlJ& oQRp{7 hoT[ AB8S5y+hM{-9JP`7Lph&4,8hk^y^M?!tMJմۦDa P\"!J׸ :o*ӵĻ+zG[Vc(<_,nޙq+ 4%އ뗼Hv)ZQ氬,ݷsJG-{ XtOd+P #4F;kh|' Y5}}YIg-!5=j, l lor˶C;[zy^u!xL$eCY4$Hޓ@ H=Ab'h|$BI $BI $BI $BI $BI $BI $BI $BI $BI $BI $BI $z$GDy4$ @EZ$PR1$" cH! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! $@H! tI ~L$!}4$ʽ'@J.z$bQ$BI $BI $BI $BI $BI $BI $BI $BI $BI $BI $BI $z$kC?D4fI h$" dI $BI $BI $BI $BI $BI $BI $BI $BI $BI $BI $BI $ t+^;zgjfWo%մu^Vg /{hw_?).eEW`pQ} O|U5?\H:l0ϿZ*ͣӝ$!RzP~B UKlFa~ѻf9ɢ3 -תO?[i tcҋ mOx۵!"èjEoNWI#AlW ^4ٗQ cgwk..Kx:89s@6$boQt#prg"ݽٺ_Fk)ˋk+cQAYR^J 0rg^+$m5`OhPCŒ B&Qz/fm8K h! jb}>M32%'eiW'>xf_z15lr~.ev/n|xQXŵw uy!UU\nx2 >ea0߾|RF);ZJ_)Y/$¼o9j z)n5#2އj )Hި2!ʖXRYΨr CB\ZD*o';IwŎ;Gw~ۯ޲Ytd 1! tn$$</I'7L.8JlUv6$sY-*xIl~Wھyv p\RBKkṚ&9(4qo t:g2/K̄*ɭ5NVGej= {}7wvt)Rr|7½OֶT-fbc}=]~5ʐy cB,Sa)VoC%7jmmliۼ.ͿZ/5dCcw!-ry |ȶz w~X\K o]_νo79}(:Y׮ukn>/׻thͪmշbGWo'ՓRӹw>9GSաߛuQ}/\;X >ZljBB섵kƵ,,?So$)Yϙg-f&hOr42ThN9fE>SC&)t":!Lg&{&g WnݰSJf~Xa!U2uH風] ߙ_A@ V%˥JR\m|ZkSqSVq QUTQ:TĹ[QI[uX/ٺk7Yz "/%xq<LI17=:,}rBW𛏜]I%]e}|*߽Iz+eFE\:fӂSųvNGq"0Ę@ o+uIB3q*zƮP-|Tp!Qyx-+Ӌ4`{͏]l^Yo4[l(3Ą% oR<蓡TDsfL,(+O&n-67< qV&W:=+EF!#f&lVmm:؝s~\ 1jw;j+n9FuXC69 xY-L" (_`NaJ5e R'a8 "P=nmp̉EE=h,ed) 1l<8L]ա_ݙ8[A eSďW{;0r] {fB~e fF=_dlXL&38ԫF.Ёad.8AqO,0rm$}J8C L5x7`8sKڣկ}~baY~FB6w,{ $DMY>\ܥ q*}h.G?m>S(Պ2bhlhct,{)9U8NC2h3%.wґJ-7H.J(i$y},GYVaYDzŌdx@2];[.whnwG5d+~^5#FU>N`]2(Tu.OˉUu}z4Ʈ?b0W+JԋϿf?8Fg)]TOJVxg|3:HrpP?KS\ zjeg.E:uo \Dbdsw=xٽW*w- |W(ٗ…N?]/,4N5͞lVZ2KWZ%Q)ń2nJ 籩-*?w~.FFqkJwm=*J( oPLz1سmL 3~ۂw%~[R֣?ib>Ċ~FwtYۋ(} I=z rB0A*dZYkp##ˁ%Bs[;;g%ۣLAT')JqɍN" ^Z[NP:psbĭ)@Kw ,ȍ _RWMǨGW^'cgeđ p&-'"dExZy*u;G:6\ѝ=?Ե{^[Lȶ.y[eil.}1J)/m< a䵁Or+)IN/|![,Ob!Smt(fʱRYRY(oŬ4$Nht7}oc"䨥Y5A6"J1P+@ZQ%zC`gAM ZL9ö}3Vi-yBC$.A 'Bkt M!).hXgSq:1L'_$þAʺM&D]3ɭ9=}TΩ!Qf9Kj;K}3e7NRYс{DI"&MM"'PU]9"υR \ĐMx )Dܰu"QOy" N'8ue]C4czmF9{8$e9dgruxޗ8<+sEjeJ`\eпx+`*DOԏͦ# kG6qdvm(RRٻZ@I5|đŵlIJ!걊uצIaR.^mXqdD[(=G;5o3Mjz‹tUl0Bh}YtZ.vedR<7"q؎v=>;dK/rxUeRwDv3oFi7?͸nC-8wõI\[gL-O^B& oۥVeWҎ>Vb~sYWDj2TKer֛u%;H)(8юZ=qz1"|4D Qy ޵6c׿" @#Hff 3Xik-Klw[^eKf"琗OZ#URҸK5g7NZwH@Lh>.2/`*,v|3׻5]*{ =1 v.-:z#t$"l|5v߇%SB*KUi\4hB 7?Zl{jYVK%5I !**-Zp|,xM'p,ߔEŲԗAŚTI\O:38.;W].Ģ[Vml@h{չ[]f>e ` K7ϥI;*ufS!,[n{Oj^]y8ޣ楒! noE Li;*.O2zwnϊ{\|36(󚼵ֽw^$i[߈+COjM7vՌUoL y *Gz%3$\ILs%j.kZ &2qK%X9 I8 ^ Т$ݤ|yjH]=D"XXy+`$T/T.>.g-g`5d8M =#׆Z 0uLΗALW2B*^2HPՙdJ B>(WQh pBP\R551 s=& f V$Ǥf\sfe*PBU.e 빱1$&AN ψ21ϔT'VknZ$`FV[vZ*+64_Wh p >ZF>ZbKWe jسX)0@_Cp]tU`NWpA{-mD"}ܔď d|3J9%m6F96)g1㛣X= ?s^/BQ ,#'R,)DX5OuB^'jI4TAFX>3ч $82#*G9KYds5In h^>X”9iyThYHaCN8#dEOݖna!񖖻wLitB,eUFU`lt@܅K$A&/m \B6QZmަހNx&)oH' ! 쓥8[[ՙr޴l>)2+@[q2p^RF ·C<\ c}^$XvF[o:(X鸷RPX Ht4'ö頵ۚtKeŚsisx85@\Sh DY 2d^&cXD}ZQm9z QNSL3Sm7JDpU0"$PF)Ɯģք茠ɘ@8;W`mW ȹ)]ghlϴso1'd kN®Ǒ88kdelHU>)&9FpsH_;zï;M߬2Ss4o=/Voޠ_CZдJ+r 冷MOx;A^B&;c,G\BlNg/]_7 dlֿGnmJ|잁z ^F+llS~f%_\ʞ.ܾ`dv(.-+OiRj~z-p-,BK)!nibW9XE'L Xk 0W`^c4D‰"hI<7(d JLXlpB{Fﵨu/tNle[U9ClU"YGHsE>o} DJhy66s :}KYm0oCa89w{pˠS~po ;_%4fұ~| E:q&(As8"2J@fVDc9j8}Y}1Kٟ=6SpHÇ8':K?X9'QjޅD&R ARLh=֛R&Mɉ#gx~X'Mdok0',j335js_I ?6  U$*^UGI ,`}\>8أGbTN㞧;ywP>)2(ip l:I'$R \k'h2 Z:|kw5ۮ5s~m0ufyhQxz~l<"Z|rk!FTK%)y%@N-u XDZf}4QJ>XLx"0l9+jE>.s<~js&<5"yG"u߁%vܢmxK*}e VS :OIHGhoϘ t@J4VFXצu"Ӣrmz$l#~>+Oì|=cس σ~ݘlz]m9s\H!L ~"), cKF?_'L=eFoIdjzWϘۦ%{a|S &zRК&s\ݚHl3rqr}9MMkg%X?I䬗8p1z4-S'}ASʹhJAu9]Og0uy=-x&~rL m;*2a;1zb&nȔgmyЌOTq^Y"%v4|GE_ydIYBg,IeԆfqA0FA۝"E )T踬wx<r,% gK)KDK#Fom^-Ч몎٤*e/J5zTCKzD/.+ W,VP@H6M}sz4A׌MT#m6,Kkv֦UZn5 Ka)m!UIH%T2:&O|~rO3JV12YMHdFzZb@;y v^F# M|dalMGݣa|Z{}zEZŌhP(KA agMw2HNH"pM ;mZ;@x u is/\ ?`=a!?1rn }ANEm)'%g1^9<GTX$ͷ%' >BR kYZ(Ebhir|uPHY"!(O'cb3V}F7m sO*@Δ.PO 1!:Lx„XEBk E'j +8Q$gAO!Z) D1.@)E/R5VSƦZL{/%qPQ[882-+qfSʼ'e z62tDkO |ARBrz!!C=%8l3$6R*&tR=cJ1_X2^ e /[6xhsn[ǽz˟l7xu;V)@Ld S>`ഔG\Ԕ4y/hC !I'אɞ pgmr'!(fJ6`d1ta]9=<]:kkvk Ƃ6 qp6ɩ&.yH&]y󉩢0 wZ4s!dўG&D%L{1yĺkknVe*UJ*;[~$':ه8<$.. Iegk6fx!)j$6Èn| tdT#ᩞaoY~eVhľǶ^#5%Ձ̌3 ʆh\9)l@f.`I$#QK$6"=RVF ,i! xe,=k޺^#^ @zq4eiޥfɶz^/jc ȭe2gf00tBx#rv`&(BBK H${zqWaocWN> [zMЬkw?cOZuh(6pbub/bQqgUޏ1GO=؃AijIj(}<peA F}&7ȭF Kn*i .:S'%Ӡa(\*FD )3HZKQ #UېI2d pZx)[ ЗzOfnϸl9 rH+\mxubQyLLk7H,?(=NOȊ62: tr L+ uL?{Y*@'͏W9&u5TֻLZȖ4KFύBE *GJoY/Kq.%.qd}?erMZv޲|&ҀRgTxq!xe-'Z2ijQ:DmY>ݥ_HvfaP/YV`hsYS>E%؋)\KI:]hePjoyG,e$0nI{ 8 Zx JDco Ww6Lȱ2rQQ#4\G90 >\0V@/*&vl;};Rwݹ]&lU98 ~tRኖ ,BUDQՃWhT?`x1bx9^Vo|3Z濪4Fg/xФjƃnuͤ\uAAufz ?`ը<] N\-I=@/o9>%suh 칿]O??sf޶Gj хr-/E3RsB bdëV,=06FUMUDŽG,< Rبfz#æP - [z 2vtX>$8 z|IϘΪF+:mu),%̊.Bp0H)l[z6ܛ 7bjT$_j WgEv7+f*7*ܒr+qj`Ǫ? uG#<gGcՓ{ӟB9"jA Go6ǬHY niAӚ1xH4mSB͈~\kH#M ?z K6rۤ}b#)_-VUPg=5(E pi%̈́/D!Us14Ok5:=&?ո5eWwN4ڔ7 Kn.GT{^T rtx7<,v;7Of6V#nmAfjuLn?pnѮ,gR&9H%d%9Dp9gTLi/OHeGs_.e2{D.K 5 (:!)3 %:v;!]wf 5&%Y.$BPiJd g #Z& +:*xY_tЮݚtZtD敱!r}` SrKLX  k L>M9C\+@˭媙u-/\X*$g1m"fJ`FE`2f͠$MHfs%EF~{Y{}Gc Σ.UVƞhINK_ҫ˿ =aV y-Ӷ |9>Sir y[=;⃟Bb`&'&9ed*u(0%9&7a¹tLIėb]2:'x(g \ILDDI;BMyЍz3Ygy)$HNvyws)* yÍ Sȵ.=2Dk$25si^w>'bL=Y>oʠ'N g)mvo<}#9h!;&ǃ M :ݎ?NA31O'ՠY&EGzW=rt%Δy(g1+ 6IJ8BpI&$Ι0Qdc\+pL@s78e~m3/8Sв|-:U U'8T,״&$l"iG߽w`x=(}_-9 9 4\ XƴVSHxo|qxU P[ tf_߸_l|6۩Mmm3A=a95Jh|y7׃GG8GbIKb-uf)Ez/iO' Y%ɰS>C\@3xAd`VK(%^m6p!qE[t: @%B^"K ͒p~Op?tkynK-^ 0&Soֽ3V~` fW~ J&۠1_~5&`KǺ=ܗ`nѬԴ{:eA%"YAY^Ľ&Em3EtÎGtn-c+vZ#;#3NAC%$/1qg0[)0k1hV %Y Qy'z rF~vnظʮgh WEj*T/ܥrÓ$yAǤ $|ܤB^2g "#yUۜ؊zWQ}$O8b`V۬tR0inړԤDD-uĊ2쭬Oʲ/1V}",% $l•.XB*g {ft&4amQkrEνO$VTs&y\E/z&4sKoUe"3w2<3OɥZ(q-PI+-RCZRGĿh8iT.?#͛3MJWmF6Tw~ln|w5>\xΖtFJf9N[p] )ī?_@W'~03uDْT[:jFt7#ug35 bQ4eJ>}hxéV[!ZzV͇tTp*#aEe,U_P#~ȫa랶R3qqjT?_}nq:j'^pق֓ pNӯT4Fs?%*hI@xTM3O*^o/? zh-<2Xm~yP@4W?ߣiW޴ir'Ap]ḹk Clz-r'D84.} eo``ВloyftpT*SR7.nJ:)5h}I^k9"yhM) rvr#@N/:לw>+|YrhiT9!g]ɘ$ xCQ~S#ރG&C9d L4 R1[ J!&zgB. b#^HH^G_f֒g^8#ŬU4>j⯋>O(Q* eԦh!+;p{=g̞3fW9c\b&jlAC1*h"bFg cj=VYXf1!Cp\J"\su(x6A螑aopg͟(/?6'Dg:|5jUc3[}q g6Gc^u)ؗl0R3tk|AĤsTÏ?cX[?. %d)J]a^%'#BQjbٻF#ɾˈ~-Ý@qYH$ˇS=R$/Q-ڻ {{Sb)DjkRšg'v E<9iTKʨSuG"y/ iMQKǽ%Ɛ Di.H頵eM8OU@y*Sc45B%0)z)FcU kET Iyth렟,J+!ANA!2J!4椌&&&DgHd-˕*U *0d9wH#n{u̝1*= &2Ky9q3i\ L2^%ZIax*-WQzc$X\3mL8s3";9KyC1\{rUFbQPDž 1HPNIψt'1fJ*SBW+]ʌ(c[ʓ*=y}_k}WiXi|DS0tq9,Ert5^jԷX)0@KnwMtU`NWBb Dڃk-mD"=b'jnJ PB%D "簈3B}{fm堦.;Ͻ;?\^Vy~χ݇R yйvD sa 9c&yMb$D"׉Z‹*1UD*&h`ĘlS($82Gb% ,8xA {ʜc SUw/ݴgnp%,ypW]1ɛVݖ`e)roi뉸&8^((STI+l@{UueLB _)OgSa{t{6|fCqiym-Ay^p48Bl :Z#1\2FWJ:a&(E_ yx~ۍ'8G}Mpw}z$_M'Lm2#S篹®Y05amkQ#[ZToV_4^~L3x;xL`]^V\ S0׍;=}2mѐئ\i8J%rP;^. ,Khޕ?TZ2 »aOr_ 0kg77HMyr-tPMjT,{a۴Ad6MJ RÁu M6;Kqꪫc=* t@:k֔zZ 1Axq!2ۊ&F&86vL~ 0й˫=klt~sU}۫| [V}Ӯ<{ږƖԶ7^d%>Yn pPKPl@ ,=!Ki N$@*˫8~ ɝe8kdelHUf6Q8}਍i EY-ڋvOt.K?~ckym։ʨCZNno43lR>KW+(a#bZsXU:V:鉃6߳ g%jvgv^ciEp &P,IZ*9kEYBڥo3>^>@ ̮n7ږœh 5mi[tqNj\"돏gp&Wݏ~̝xV x^JehRqkr-?_E"ڂ`pqA?~aij3VӲxq% pIMBS8BTTZ0YS N $Y)Azl}8ͩ8?> |>ى=t^̾.a#!û6.C?{#F|cTpR,.ʏZh…k,Huh+ӡ^x'K,Ηi{%/Ӧ"2(Ebhi>oruPHY"!(Orpk1Yx8zFR4MrV =@:?Af>u~hBe8BUI7to'd~N1}w_oᤙ֩<򧷣 gAipZT]ZF Qluc%zYRl6 INN3,qbM8&$!sMcŒTSOH bax;?k;;!&bub&2՝Jlk Gu4oK˜8"zLr~+Rc0;[,GWC}bɑcUga's+%ܓgUNӤ7S6K[to֡5!w6zF6{Ov5Q Sv N"5MfwLuXENcǛKdgt0w7?9m_QϦ= 7C/n׫.[w7vkCSs2e_Q?C_3-Lqo`M\>twm57%$UlնB&AJb[u I,(g~*Y;da#A2kY &2qK%X9 I8 9\pahQI@ZXh?Z*:E DK VH rϳf9_mAuh~/!s Jg0L]0xeUC׸T5\zv\G WBsY%"fдRZTR~9'??gw|/P"8w,F)qsj 9 t&"&䞙Ȁ 1Pbb#ZdvRpC)I [vŢO,Γckz6ta] f?Y9/~I%u Ĺv pF3"|T!3d8 D"m䄲V(Y%G.6'J,W=?9c<]ϖ`*[6#imC-:b%Ͻg!XL) ^aPK+dKk,-,A%WwiSM^q(&-–r4<`a–yD:υfB)j#FiYm\D"FZm P@_u$=À#?˦t/;aMϾmy-+-[\tMRr@ ՊX^;K PiJE/!zFTM .𥹎T;)e֖9{2xdՌy4"V*ԈhQ"k?YT}'T^ m6kΞLYzH6rLp#R $ʘwS<\[^IsFNLKk-M(&L#q"&23~rZyN%ATVrqORpI8\;S0JLHQAДJmv1!tl?vԏ[ wiP\/q~h+s; p}>?)#Z7 0cKWJsIa\<>u ud\Rd ;er.GJI)v27p9}/+|_!;\HrA&n1,֍,9<篿bel@_]_UMl,y #c)xz:(Zt6d?'t_O3 G~4Mgc:VtD;V}e(NQ#d8śxV4/RޠSGB%Z/0GWԵٺ~[=}&8a2]LOd|x[.vAB6ty{;IohzGuH}Èa@I,1Mg*Ժ{c|ы1ũQ>z}Ur.tQ§ҚGnTK\H[)ߏ~>9^R|} RBRq)pl~.ݬ$lY&Jj(OIO>|?~_~?|\oû#]z/KB)x~{|miPs݆a\j.g+T[b C< TZ򦍱FW; ҕ+P$Ev183|(o$  dx5OgF$#|.M=m߃w,GKLeZ)]J䈑gΒ,g !uWN-x;>y<}sV ;\ݖ[AlhRE¥*ykP JF*/zK^yJjˍ"ըr4NMtbtL#ρ$3G u+/3O/K+AeftoOJ0q~2L*=:RFKE}*8{Bx(6Y;)ƢMJRD9̇l|_B8"\J8F BP %Bz &E  =gY2Wĭ+^4|+Z9%N DCdLa\ BESbUQ 6a: 3뼓1͜ in*\ զs_.Yb ;`JI. >,d)9o2)cKH$,`!.fOBQ:sM@J!@rLӒZ?ggu9ij 8$3@UP$Q7(m@`l!K1 q̒W Fւw8+ B[JS%ɴ%ߖ2D)%r,d>A& 93Y,50)3(sEpBF B+/G7+Hjӹ/PWw-Cu wl`<>CD\cZjo+7s>z%HmS,hu9$N$BnA0 H`% A{u ^&[\0L9PD|*u4 kTCCiO@W/ӂ|<[E>ZA-wd̈6/-OtJF|LKbBe2Q$gPI{EѺA<+z$O*7i 7L*OiCElygA)V&f`$8Yr|H@& &@)NUV}3y$X)vZwz֩~96}ѐKzc~qlvG.KRz3fqXcyJ]t:}- xw)t&&B2m;;5.njݭtM'Nlgw w<87a<]W٠M n[:6x7z,`"h(&ğohmw]ܛvW6 2cPW/=/de2JK5u ]/,F u_-otyfBN!&"qtv%+?~6sZկͿi*sTHJ6 g][tkf* 3A*#aenTėi|F'b%2ktcfQGHHz5[Qv@5L/ tm*i=rҭbw蓹_eE'/id*@,4C$`%Xha~Y!ߣ e{@5fEeSE*,$&aDJ t/t֞~xɚiY**z^Bav׻ 3>>P?snE@kŸ}p3C`1drz86;5Q֨.Xy8"s iא6!f7Wp^Β^' F9T2Z[Q"w<`͙+-8ƵY#3n:n/)xSn'ȽYΦ47%gysxmjAPژ][x\A ?io9H`Z` 4&p<4ΫH ,۷jOe2jG\,l4e*ĢAJ<ϚY0&P{MJ>%K$|.iDɜ&9x'SA te`1VG^1WryC(eVceytm\%T$|d0LG+J\ƪG5q@v1l%x $FH%NS` 2V팺JB yO"Q-:yE"U"S)7.c~@4l [s =krQU*~ zٜhxё&6rF?M[uB[sx0G'EKę>Iis>-MvXT(㡲PE‹-O 2zغשq]/rUO&x|deKtjebt9F3hc $ĖVfm(ƞGdB"66QCЀF84l;f]8c,[;tna6 s_v5= ح ΃u)c6:CZܲ#d!kZbr'UJr_2[%J312Jr)I ee\md887{o ]is4-/ŭ_tLĝcf)2BOҗ⥳Z2BD d`[x~ѴC?vC>w#o ]L(EX]eV? nb~ -"Vks=RޏRŒrASXv[G* NjdH.s./ re22H>y.d^3$HFF"N-.D%S[(V風Hb&8lAx)\x#$BW&-AbHlyx\m3ğޣ2BWwrNOܪ*M%E&WD4RM^s36 Fd!h~=nFKY7z\bH\D-R(pk*=287[Zxü 8_{&9E0yŖY~!fY^.n7p#0䤎JfZŔm}!։o41afy+yY|4>gvg_φ69ykX̆2 xY,eAdS7{1%7Xω:;w%sP`J[SM>*Sd" SFt,@r,qF( _4&I^"rG4D\S̄ E%䁢B[NwX@"4 ȣA0۴ajPhPjny: o_.$!]@g+(Ztub}J^Q8ǡ/K.v Û> }qxJEٍiϞkˮVs]9vy}R֦Be2Ry,9?џ5q`s 508}#mˡ }dJ+'8Z6ƛTbB{KD 9WJ[R72B@ą`Y*^S2:T .(& rl=gy9s)nBK1F;K L'E{H|!A>1Qײ)*\ɐoJef4ڥ Y-5if疢NZb gK ʡmCрUQ٘EG ,%k(] F$J0Y+`ŦR:lm Ptc䷵;ڟxjl5 ;856'&FSc_hj0//N[{=kab$u̸VQm"3a%wI*Ċ~+JgEf 4ً XVH3ZEhEIuI[cܪ"MYUėX0s`瀇 Baȵ(A.287+qV8z؂r-#M&]^9狕k6}tC/yw[Ru$K:,1)&n*XY69P j Ό(ƻ@oE,ަY/lCLbԍ\o|QFae5Լ60qa)LjI˻ _϶N'4f-V^l@E;ZTDXQ]U櫵仩U_gԈ|qQR![u3M PfEh[;#ιTD k,e9ĹyN zeݩ܇m2}@vJ |۳K#H;0fz>mU[_Ou]浃`7]r=j^w]l?~v~2]HonjKPLi~8)T?]N /~8~>x^aC>+vgXhmgf@ӿMΠ˞ZtI A-djl7.U/t\ ڠYq*fZ6n?` tN!:qGv9/s)KAJ]!`5 ZVwyr+cc5٠4<] $u/UWhd!9%z}פћ9.PA^ig=dWc"WSE:YEuYLF\ꮍr^w$Y^фN4暰!,R&^U,6}Ý1\ Ԝo/dtV~N1#ACc.:2zIƜ+E\5Je:MR}"餮1ХŸrآko( 22Ii_YX6 ՒPth3a8dYR(R]RS݁Y`fb2*2"Nf&'z|eM=B$q6yD-̤Tx (ƋAC4UڣJ7A**Zt)\]h)7TQíɀ< ;l_jTc#xPr ($bDN W$2+Bi,Ѹq / Cu+S{H %nEf0mP8ڂ?3ΎZɢ*lZϓW14+JXā))ಒw2<}XQ?UvD{Ң"Q1@>bQV]kPRDh7P1!5K6yBe 1-p7<3J!t92Vy=jͨnRD#+&K ̃?8Zi2#s4pqإ6FuF #%0`,paE^Y` E*C DYQ#ǒv7XVHG}x4YF%Ghhj޼j@R\ ^@VAcQZz7*-( 3]$ke akC:[j:8l{[`tZLi۴M:lbi$  xZ`8DL[έS[=Xiow kH繨 T/HY5Xhx9zS~ xoC{o>jk>sp0)Q^"o*Js1#%pclrI,Bј2!SAt GRDYZ# B\ JX 7nk`XhlFT(>^.W؊"D W}\'A 0Fu/x 0S! !,JFJQ 5z8YXRuLc-x;W[iâHY=zV9E6H0s#SÕZہXf=TpkD%ogFY,JȦa]JҸh µM.kX#9kP*my mho:k/5s%* Xk m@Ru2rTJ(Fvv0Or;ǨzFvf΁PcH av޼J]kq՝ڦ|٣ws_O49b*mxbI[UDkrԔ}Ur/T-ו&I4yMյJ:cО?WG(1~!DC~Yzah_ zcޔPV#s5{3BiPk$ y޶G_jb3?[.vĉԹNP*oPыMM[~AxOϴ3m^ VĔOB /k19G_XA7aU$eTU%gU[gLM4WO-6 a2HGc,eaOqQB2Q(F c48AF3 BgdI)JR!̘ ,{at.d⃭}븳52LjVt-NP/Ym*:j׹*9_,ZK S k{ x;v2LsyRhep]>7:l=,qv}sUVwCXESQGQسهܧXz eFCPϥ ;;{m<> ⻏ɿC~n~fw=3ַ<tG|hEi߮S7E^D۳-qZq=qsMk\8Ĺ&5qsMk\8Ĺ&5qsMk\8Ĺ&5qsMk\8Ĺ&5qsMk\8Ĺ&5qsMk\8/Z/Fb\ROn>E\ʼSWS}9{1[L'Vt 1!J I/a/~>A?á;{o߹(?v'U8UD ~C5l:*F>u}f͜f}N37-7Kn?v}9OAk$p{f$H¯VXޒL[xoGnJff`V'/7+-'Qnp#Ɨp7 3[IZt9 XM]PZ,Z9>:%y킐3z,ZdM(mJi>ec6!?,zénCj{f~{YM]c(`G}> ;5}Wś׋X@Oq.FL6P cd^DlxeJ,J&O1xw$y|mgy O-zAo|g5v.?.bӮV\iD:ND:ND:ND:ND:ND:ND:ND:ND:ND:ND:ND:ND:ND:ND:ND:ND:ND:ND:ND:ND:ND:ND:N߯EIޒjzގ͈̕ZO^̠:Ob߃c&F1b@"&1 DL b@"&1 DL b@"&1 DL b@"&1 DL b@"&1 DL b@"&1&k/juXVspߩXݟ| SDj]zFp!o]!5kKC=8"j:P?)8n댶'Ϯ\u/퓡jwO`]_7|b\ܼ@;>qpezy[918"S2A?@M)6tE[73jM8X b)//?>ڔ (C*8MƤsCN}1sE,UC5,`EF3瀓nc E||huSL?.Pɧ˧ {>vô>l>V?禦g.$7xfCL 8KW1]XDbDy <2S =yZ}t}kI709k]7{D`ByF<\*H _9rWj"+˲QYxJpqR ūZ U_Qf:Mczʨ$>m .l#Gm.||-f}xNJVK*\t,y\2&bS8K-W"5(2L:{*&ˍx:Nùڣ=vUl~h`7!?|v}!]5ɵbv>5oUi]޷}'`Z6-;gܴnGKCZx6i_/L_A`˘.ͮQl+߼8O|^Wa#?oí{Y7#N7n20W"nL{O+եgo&p5F?![;x<6Y!?Xj~{;ޙhw}[㡁[]e}ZGL],M`}74 ݆F4#ʡ:ևZkz5n xŷ %^KF3q#q&Czj=uV6]ʮs?}ntyuC?6gY -_r_2 ,[߻MZ]5 nσY|aVsBc n]lxḃ׼ܝf>q6)M']Lfe5ѡ|ceXl1Y~:Wxmu-~gדGO(IBC~n~V}׮#?;nKJ~W]_-Dtg*Ŗ/jtڑFTJ%վJ\I%3J/|_G(JJ"fw).J:ÙuRKJ-%gɼdUCCw6˦p+,CcyYJ4AV;V ъDhƠil2y[\_ulׯxׯuE`u4Iy/ /VmM m]2[=]Uo4w`exu7%]_9HgG}K+W/o0#F6x{?uˣջƝT5)\tt-j]wU4_%]ykƃvWGQV{W޵q$28ّ>w>o?,#IC_C8$M%ڦLWU9jkU8"s&QSBLX B@ fF8~O#rWQֿ&=Bڷ lG8և {;œn)+{8'چ9mzNJV>kcEh)[nhm&RdS(*ZʻSbyiPʑH&Ϙc IBvL@O$&k:%E=*Fj80:g4J@lr0kVpaa1rL#Vep!{PUsĭs4!hpi ZP*Q*"s_[Z zAA-FfA翨'K$gޛR|k (þ.traV%ò_ȓzxo:ٟxxKIx^T+o$qjA[(+,)k hB Z{ S ϥV)oC}B2N[D-6*Y3*ta1Uʃ.tᒢlYF=;mse?4ӯ\c[8ZDeCNK)Y$E e佀ĥ7Ew"jlY8 سe4M&<Wж#$MF\n+]Xc#gvQeb.AkZRh"&joh`$M\,Ff}]YV{KǮQֈqh5'YH ZI0&@dLDSXl^8HNh 2)<(AGKrŬs[aX-/"jz퓯Mazקl.քDTt$J<3)`ʀR*9u 2d(9bܱ/? ې2u]#7zm/8%< "DяZ)_> Ԏ_9ONǓ{׃}0Hf$RsTNQ XG'48U6?n{ lL`oa~b#3R9~O"Oje`sc:l|}9i9}g3m'BzYfz%_.{teX5t .RW4* ; PF+r|.g*mġ>nhW 㥩O* 1>hUtIљ3kƒpl6g7:?˔"bngR}P׈%ۍƹg~}MVRҙ4wvݛB)3|^? \_Ɇ~~9UٴͻOJ!uX#s`ϪcmEM8=Z;W̆M $~CBݩ>oM8V "}$V(q :Q I_KQ$)-d1el)BѝA>A־᭱Cwkm孱5#'Cf[Z'w-hL) Vk4jjj^rO|#[`-FxrG1W QTZm̛_RinٻhVHKmz}r]Oz^ݢ]3\t>lIIQ=4pK%Ax% V8fͲ}2੒*%X\ĽJ I.*GEA,KZZ{>VI@>v_i#8`Qra'Ah~kcCjelϳ'$9õ⢍:ϑ93-ZWٽn_w^7^̦ /op̒0|n[IlEzyaGɽ眐pzGbHþah0\sY^ 8: g0b^ǣn1tÉQ)Y?j}=+ang_b'#y`xs׳& >c1:#Z?KviËQ렙t.Z_pvAy)L< U9O.hUyT#Fە\]g?/~Wg/~|}Tϗg_4\[a}' x0^CRCxL-S/&|q)=t\j@=ӈ!vwh$YOC2IAh2i˾qƴm`$[y8IpKJ죋m+ ip.kͻkb*!TQ ;&U"Ye@2 2yEEKƠ%?ӑq,SF-Wm0A=-ZPn63"V5}!_nؗ&KރL_Hq˚dTY[v] kIk(٬YuFq oc[ϟ[5:]pef!eWmnMoo|3=%kv56cuocz}axȾ$Z?D6>Cv籱x[M$HΉUJjhz.%MRv.C}-$G6Z3o[&21 Ej,󌧠 #>MRR-6I&c|0jL Iv￿[MRՔ(aؖͮܥIYz%JU(*3&g$T$2\謏@FQզ:ÅO۪f]ǖBZiF`5[qܠ] dž`#uDWALWZf{4 }6װ%F(}D xiё9u ʘ90`)s*)vJdDL2y%RdVlEOI\(rEvP/ld6s?dfQ+"|xЖxk8|LeȣgCLT$(leQ?tO'S܅;M 4R0Z]^AKAÏnnPҿ^%4b pv%{j>`fr4-WdOԩGdԿ}#=@ -y߫wkT԰s8_԰!ޥ)-j34Rd:=v-; +VCD6N ^Y𞯫hͼ +9fee7bsK^7zﮚ?z8ofU3nx0 ի?T~e6u%KpapY}K]>7*\}0ſIC!y%eѳ+!'"L8 ׃zӮgO۳E'ۨ% ޓK#ZAS;,eG@E/\#A&%^z^D~l:Kzϥ\%ݧ1kl`fYLUnHڭܫgg]Ϟ5rԦ3ޫd]'阕$`bgs*0j%!,4) iԀ OW|g"sAiPlg9{ 9,,NQ3|HL.=.5;(jaͿ,YiX]\ܕѷkRYߌ[&qP)&.0AC7C򘢽o~/4Vk! ^?/|JǮ^}Ћ/{7/.i4Jxj2^M]yNT:C(Sj.蜵2"fKn {z NP_{EUbqbz|ō'8jҥu{]^ץu{]^ץu{]^R=S7,֍u"y8SY '#E-X"-+kH;+ܭ,B9T]V ^pSDe6`5WaK灁E-F-g4I.JP+tO/)Dd,c.2iJ2 N˲eD85 4h2/wҋNI4ǩCH+s/{z&D.\ +s#,TPs"T右 YY$ÿ$t[9I_M `YPpYМ8|L*jR$9Mv'Q5a=YII/;%g :H34DSr5j\]jEw7y?UC?µ|)o{8zMu;(C.M:$4NT3ߏw}LȅՋ#z׷UOPxSmϤd%3h  2A^J&¥LҍLqL}KfEY9X\4yԁjWPh\mEiCy3\֐S5rB2{Kf_[zbɢ5pZ=Vѫir ?<)M_oPI o6ۋ{ T?7W/vz=*8ۦmԶvnW$8?d_]R> ^={, nvwjg11 U(gAEOenz}KtF*uW[vfp\C6j|.Y5lѶ=T刨a~/,VB~A7 Clu(U?Li{"O(.uOۺ՚n4*HЮό#hPgcr;~֢OU]n*a+uQq; * ǔH:hHYz%JU(*3&g$TSe湐Y>P8e%)*[vVͺ|N1Xy8J{62V.؊=6L}? xS&Z b:DxfNCy5!ծ -_9!"NuIIU29_ ;. +KnEN*Fe9$B$ 4:!gLlIG~XM;99fB;= N!(ƙhcdb"Qܟ4e< bB٩Z ÂΙT<ʋ(e)樍uVJ&3I1 Xl΁h: yN`{ jGJfk *srJsI &qInb[3/tmO0ilRe椇YʘA[,{MY$)[ c*; +YIMpRǐGÕO|`rnfcKxg:94WܶI=.YZ-86̈V@F@ *pb] !i Y245e),sFq9ni]'*R2a]4KAB Yi`YNk˵Td2k:w:\ DET0 )3!z]|"i,5kPŝyȨh7@c4-P%.QȽ!c wFаaS]c@;z4䤌֊" ǙsI#R\:KPYݘY[;S%B}_ͩYvk}y8Ľч= WRA2&H*Ԑ)%as04|r:3GEiB:eYڲ0DZp\Iqx=b<;`.mEV'ukg ev`{wstGnctIѺvRX3ܻb O, @J8 G6h?q`I''N@H@1Ʉ(m~W[5 4>ogE^2g(yǙg/z_>u%eD,|ySm? OUXX}NpNoK9N뛾`a#}"!ܕπ8FC+ ͔!<g.fRmc("v2pցUCYʺev]~@f!'BZb5Ѵ q}|/iL?mEi X<'4^v!K?B~lh 5&(MJ qr )QɕAXf*+?: Lg.V8 |˧ 9 q1ȝMp29:ruQA?"Q`}VJf7o7d̙Bj2q\P|$#;ZD^EtIeaWtg^8©tU8;=/.RRFQ~lIAKAfu4ArX2 UW |v8;?ۛJxcJ>Jq½ǪoM4! E/}|gWqY/jzKݭԑ@ULhZ$d:JY|ѺRk{{[{_ Zؗ_򝭏T-6ykXy"&dZKn $R LpnyIJRX"L> K<gc'Ȝ2ZAʪjT:ȚBM#w: 4a2/77)㥯Ȼ @^w: {~x3*10ER\}6fy9&kT3-)!or9N$I72I#^T*{QL.ǽ5Po8aQ뎚LIEOƘGM/MH8΁4J^7 iPJiy>.o??'6*p.lE-H5x>ڻpQGq!cL@%Q:r9eMp0;. +bac13k8؃8e>\ [ľ}, կPܣ$˞ݙGBz5w7'%B?w5H8,_C>_nxLfտ5ͻ=JQoi ܍q 6dWi0K/K{'mgo9}ck?^^>#ng =XTmurTZ9FcI9 ԽHʢ2ђ@+i* 9s_WoddJ턶D.,5%IŒ"JYNXOg*պeڅPR;, 8F6)z@cDMQI+6HJp(]nf}ɱN3m]L8Et–yd#:l!{nCzW/w1˽1YG -EJK|N DG"mYAHƦd*zt߼=G5a봭r"h 9ε{j^+a֤:f\06JZÙ C)_~k gFfA ,#gV;ъa[a+APӕqr 7NIoR'>.Q`u4>XncRxU {ʬYŅbВٌΌ:o]ho$xa_mr~5lxKUy`|!Ɠ;!&#\Z6)R{6:?ے_J头o9`[**1F}EqmQ"yLyλXTg/UXgH_CFY&m $(GhRVh.Op̨]Y2/iMc%i{\oxE?Wqb ǭψ~MzEӳ>?*nuIwɯYzeO|*4Ŵv[U˨z)sɞjU0%YkOXA3F@h[kJVT؜`VlnW\Yf').ȊspȒQJφYLaPek3Uw-b@4: \tF%=1REfDd!e:CX;[M< {y(vtYwPd*U#cuxZ\D:%ON}mU I@ uv>k^>8D$!Bju2* jlyvJj)NK֬;5]M.Z|k.]9ΧdJIV_y'g}8%1ZOϓ-Y{_KQhTG΀ !YH9r# AKe$98m[8rxA&K$G嬂E0V!HFjٌtjXX3vBs¯ {T6;/~powե٨[?m!;voL(3+Tt ":ΕAS<+`2qdɬh@046+ Qyfպd +&澠v56>hr< wL[Dp[]J$ iFrbN*腉ـHΙr21YRtj "+:I5P6 .Rd}`!BHTe<֝xK{cWD񀈏qV/Ufdau$pAeqfVf棕 yJ'nP+}6KJ3br n%Ȓ`J2"V'H:.v*YKvE_x}ȝc*gf12BO쥳Z2Br!Q" Qj2 ѐp9pq_w  lCy9~ܸUũnNp/ȍcH'ЯT(;!dܞO5r\[TN"BӓYv[* dzd!K]G;xA~xs.fmI:oUaJ1ytN,2h+ژVfR L?{u$u.o.!\ NPd Z!t22PjTf ~PL'Y^W+_q- ʇmޮ8X_ƲtVyƮ-Ѥ/zkrE&;Pw7irY5pU.֓NiFyDjbF`7{ SFB?+~x6#YQ8+` 2'G<-Q] oG+HFG$`;p Z"._!JP45im$&g5=]_UWRRR;ԃjV}L/`ھ} 8Ch4^vB|_r-5w`V=Ds K/ ?<^\]{w?IVNbBa=m\J}mX9'QzFk >C0ɄDޤ;٤)905&;{{!"b)5>ˏA( f2vUf0Wzk v36ۜf}o7GX7( HSge&/|V}L] 6Ճ]SC([W Q!\-pBZW j!\-_Pť.)pBZW ӅpDBZW j!\-pB2XH j!\-e.,pBZfBZW j!\-3b!\-pyBZW j!\-_LPDBZ j!\-pBZW j!\-ŲٲTFnY^ޫu\Q,KpNY4*MD8.ݖ姳#}dC{ew=:~;~Ya'S0x$qA@1hBȄv<˂̇J 58I2ZDqEK@I 1jM؝h-l $gK'6O49 .%2Ō :Gx{OJ"lL-;BnV"[*A9ۜVu|?e`0yQj_;աիٟ]T ;_}W5z\Q…Z^p?0͟ ~}Rw:x8/Wy%98+6ml?rжLsf=: 8ĚthV܎cW e?sv޵emG>fmO]!F9fhuӲ<ÌD#[^{SAd0Tml pTodT͛ea dck w{]PZ;1.! >cUW5&\_\_L6>\ 0r3N}gv0Ջ%qևEu5 /5,̐3xH,/R[2NcrZLڦasva<3_& 1>h`BtIt3k׌?i,)"<^Npx]%ۍƙxm^_KVZ0KWVU ɮuJB7`)>*/߲U3xfUs:hGM廞 şo4%VєѷNϜ)եaìo{T>̎d_ 6["7q|U՝Q +%;HoZ^7ܻcXNg^6@fm >9 5ƨ`SIg/RP`#)Q/n$hrP)c$pPր;m<8v(CLQkAĺCszg-޸ +7)RrZE@Kxs QH4Y4!r|hk)W ܆ۨ $72BBB5 bjA :YO9T /6fZ%ka1$UyV/)w֫v`9_~u$@,^Ig>gQ $ iTɤ|J| cc4LA׾᭱Csk m孱 #ek3mE4[cKמ6)%kC yGkc)μri#F߷D'q'Y0&X8J,ژ7`ٻXVHFlZpྍ\9ImkkGmʥdV=Ri&rE/yP+A'8;mu;$.gJFɃB JJSV<4ZT(@{cq6iulF2KT2&Q \PӃע ]qOSP l0H) ǷV;GA$:zEutfEU]O)Gb"NR5˼?B h@X$Hw?VSkc4;}=E5zLbTKsON -q>ʾfqp1ip#`&uqdz0:CKѬ aPNX8$#{$䍣֓3wP(׋ h J1ĆVjS xGĿVQ'+y:<Ӽt5+kuS=yx28_x 0. snltַ\튃$GR6h"==ec %Ζ@o鼫 fV Tv̽Wâ6烋M'[edsjY9K|%.t$7,p<\9.{!F1*rD03N[-oOA3۩Y]삾KV(a'hrnu~^ōu. *q>{pv)*__g?xr//< eߞx|/th]"*; ߟM뾚Ms [4-ϧw*#wvFȞF )^//˝zׅ&>NOzoN BFΒH(κ[H\+&m1DBT`)OtR}ty&!% k)4ךw1$ ĨUDIddø0e_}&^Lۘ؉9ns1w:y.zsVНCw;PPx)Kj]_ "Õ j]Qmd>B!'/tikva)ީdsYaK"Zj."yeKIytP*@v4h.'ҜߌY˵pAb"&Є1AxG4i\vt}B ȄڶE j,خL`'}s;gYG<{ n n] :ڄCG\Rmgm4q)g४5.uF 7ڨ]l_wdHlI"PI<ND&03o[~aoA?kIZ GQQiւ[SdKrIRޫEjp3k~?Ybz1mQOBmy4࢓ Rj0䨙LNi t#tzbCz{E;I9dO3nj-3R y14B>92!Ҵ[yjh_vg (ՇtV>S{/LE9ˀ5d dHg3yE+`!E+ϻ#9BY8+8;V-y {Z98 JZ}n^ny}5YdDJ֤sHvX)k˪r)Fњn65εѨnSغ-{,]cszVzs؝Jen92p{KXrmj7u/ ^y%5<2}sw}#.y1U+9×|gt7e,E X_7vY>P>{^ط?K:SF~FI3T _Gsb K?P;>7A8|qͩ )TGbH Aznl $ ! O@xF Lx*zWcjz|fQ`i>Nd5k\hMun%@@&sb PsSJ/Ʉ5$+J9U) 9~3\GcuuZ;=] n sm'ptH!s 2b$IL> Q'u×*H\9AcoCq2 Τ uQ娀E%KGS8{nGYp40 fx2DxDnyb$$*!Rt B_BlZ^reyrӨ Q!8'E(CK1>T\#+qr8rYl/ǧ$%)ˡ,k߯zHJFTS _U*냑v|Py!4Bm7ĥþ0&Ȁ}Mt^by %SXBZ'!P D;-dwA%c:PtNoG^hl‚q[ h"BsAK^۞$U@1T1(Ԡ1 A+LY5ud6`dt19诤кvg 9P9` N(4uNW(aX'Rs}-Uu|d\r&yJ?e8Ǹ'r8.&.bկ2~F5ݏ. /~!G]yuo>-ړ_f]~C>Vi<6h*/qG*q[no]K\L>M#vQnԷ;ngcܫUSx^F mnmhom iZ _j4[dkju%<+QNK'*V*w$_*ST,\5sװV($+3"N4ALyV;riAuba}עQ"~.!x gz$eޑJ>{u_l# Pk^?J\Ȕ(q4ՎqR+m>JECS:뽨Tu[+&DknU[N;RؼR~K>5wΛɲwJ1X=ʡ͗dF\y&erít^]A(#+g#-KL2ar8x %}> *)Պ&0A7rbIDjU/[LJ>5?F[砼  4tlJ2*Tk}R],F6qZ5n~zdOȣrZ!\5t;,f:k rU=_&u8"R@s>OI3x`}=?eVgZE槝^]C*)b{|+U{յ {/kZ%CTKiYpT}tZr(fq14Kz T?~Z\]ּPP OUFHB,l>5V[L?72Eau""[{- ie+[/r1p{i*غVFq5shvdهsGjs>)kOşwkuJeqkMqS#Cc9_^eyY[٠^Dxt[8w_Rzo{XOSi 70?%ۊ=|Ca{O嵜ibֳݾd{wU52֫/\fH"Қ V4roA@DFcA,Jrk |Jw9+w'>ޓ!jV'֗wYf~^z~,i</s `VRpr’<اR !Z0EcBΜVA6k0rl5E-TRFJ<,J1CS֓F&IsέI)(!<(QECu.F%3xCsc{K%`uhRYL%.֝wBݮX cx5kN\B)\hjq2& X"WPTP "]B H6D ϒO a£h%7FP:x Xwe*zji盪n~ hKɾp -Q_1lCUuaUTThB@K ͝"ˆX;#Yĥ:31:qɾh pq"R":z-cN@)Yn˷3J#h@FB0!pXw @e v|*_q:eE#n~|GkL~t q@7xt~r A( M'Z & W{ MmABa{؁ K0H puJ&AWAȬHfQ1h) t2 |QbA)ML DL4L%ʬbN G 8fnaXwzAν l>ruD1{;" ` ;*]3ty],ΖJ&_^)=YA4yy^f%>20j~hUIH%DwV!\PB]™hepلO%EGfL U J<,Y${Oqc_AyM0|!͔W^<,{Ш$Dڦ|MmYR8fK ݥ.w׬簗)jE<ep|ًHɩġtZ/(gh#@'$knu\zaiJh 4tzb t-HN;o.Jli$zzm$P!!q`rw$@yܡ֙UjVg'5gU\60ż)Tu^bV$u~4U煛Lg/jfPQ¥?~ U3qVy߶:ڠ[~zzܓo_bWfmMܑJ-) -5'w ?v^WJ[ʝㄵ-ݤo|V&Trd޵q$F~X&fpYAĵD*G_!R")YI= lVwLWW I@YNJ:RrI8jFfkFUBɐ4hsgt0.F`6?h{Y՘0rQp;=20D1P==AX.now1\>qwBArâ٦d,K:{cr*/ŭNpp!NT=,ՁNy8x0d f~mTNuAIKOĢx9=c)xl[otr\%eGU/߱ jQu'sQGŌy~Uzzی$ Hmӱs&3X$ C@R|P nq`9"!~kH[o:e^ 0XK1u"uә8 !ޥySѿBM[- !s&Y%RFblk#tAWf(,mޚSo1c:fᕱ!->dzW"h5dNTzNWc"T"EFT=3QŢB:XVbb)푎M0e(`#t+-[CcV  Yz2vrB\CB6=6o4iTS.C*޳#>κ+0 =4İdB&9%%`=`X_~pR|Y^g2ҢrqpHVNyʷK)C yF[ҩ #g{`Qƙh hŃ+o';Go_@a5Gw3 lMoŠ;PZ oo=w N=1~xE c=}Q6*Jl(jöMxLš""yKKe޸ *!s>%!S_+R$2A.aT:(GX@c(v}%ID-P+VY0&b&VUd,B /͙pl(Bsj:e0*a|fiBzh_1Pzk%YN4lųX]'oX!ݬ W I& 1 )Z!<VuZ7jZmyNÊ{n+Nc(wu6l2d7-Ux>OyMN-GUK=ZȝMǬm@լueu O,b} n￶nfwwϏs4e&0~mܲ wݝU,_ޣ畖C=Źݱ{Ϝwz;!r;˩Ex.cYPyEXZ?LiisOi]MdKDߌ]iHHĆ4o,/Wٻzi҅Xrܩ>Fc.AccD*:EtZt lm ! lv@w!'BV{))r7Iڕ8wdNqy~~+n^ VSg5ηܞ5GS,N bzBxFoд5;b):eQ73cآ,4&Ҥ}!FsF/@^f+HaĤre-KSNL.>am..ڠJL K!W0*/zrЧ m/n'gM~}}n,HFۜ 2NgPBDV>0wq)BHJYhQ˫}uٜxP)j]<3>qs2T5TsIs#++\JQ""M!z 9XY?HʂPh !K>T3lɕkSK.Xɔ̕]:6DV['-WӞfyh&vt-eo:>j-Y T ړFϬFAJ]ew\o7iW\A4:KcHƲyMaOudȞ!kTU ;p3I>ɣ*p浇0zmLT|!BJI$t=V$] \$Im iK2($UEs`M yؤ@has4`s[EzC -R)H,(@J1 ůSIVZ!fCɂ$y<Iݤeٗ܁F»(<\jC=?$<ɕ?||[-ٿU=LL̷] AaK8g1\$ :j('Ey%@![jFſVgV.[A(͛ŵB<;O2gU#[=N*ZbV7o1`6ɿ]^qGיChvOփY^㏭ٕ \]οx 6CdRsa8*gÓٳEwlW$ǫN^I-#c:#oFnaV'VĔGvƣ=sx>usjRuqFv]?iM +>ߍ'v}5͕zY~զN-vx47tF 8ꂾ^KfH:k?/㼋[\*ycv /NYI/[~/?7߾9i_7{\6VI\gw΃;_wM/xaU2;u>S/eC(urY 3^e8omC$TMЕKeП_*+|Zm(Ua(DleI%@! !Y6Zg8)CBN,C IM SXNN0Rؤ Ĺ B;| er<ߺBlS iÛsh{tBȁ4=9vg_Qhgͧ&g1U>sHq6%|)Ta: UMoDfŌEFTH(gȒC6Ǩwu\Muq˫ B$[KSeZʷ ed]VOTuRɼ_;Ub0TP@& I\]SчA2)Zy:PSDY<][s/QX}\cSV=^itGnn _]S.tfEe ٮ$ βmno痩PIvV7XBiDhc吕L},w5ݛ;%fj":Y7E_.'2?0~Nt,5pU qsiɳƊ=bSjiVYTۺ*[xh) %NEoSxJO=AOoLu[h~M2a~kQ1р+>5diol$KJȔ 4Ԧ(,O>[幟ADu2$}:hk3)ŊrC$ >3?B?Άaw(brn O XgLޟ{8%_($Փ9Seߔ6/JF[G"~Exn~a[d& >f-11g"r]rG~73dY#Y^;NXvLz:I@@ɜF @IR)t*Ȭ/yElu$qJFDdC+ )IyBs\^=/ "y_E>lnsk77S;d4PO˺~̂Ǜ\s%󔚄H H%"t3OC'zEO'y>7s_TUA`ZbBB3gX# gg 1u-MV=:g8a- <h"ǹQ%<Y^Ct)Jtw\ן`bYC2Eԣ@Qۚ]&_"[KmKd?g.iX|t(b[RRR4Q7^ !N?^u\|tKQgr"*( ed!wq'|g?>]?f7:7ÏV tя3bݧCCl>p֚g{5q7c>MM2_ۖqSWj&SU^q3jq|vًZ|[^s(׫/)@r<sC+LJ^My/Mc;a8 )ߔN,x?[ 㬿4YM&?7TSzOeb&񮟘Z\uBQ mo_Lvy|ƽ4جk/>1I~+]o0ϱ0A[7g,h 1obZ6ѓVݖj|aģN&%?цUjr&<dIzrdP6Ima&9$: Q Cԝ#ļ-ҕY _9{PʱPlqj-O1X* ۀ&^D0?Uz}*'qz;:m퍃*?t:qj[ą'/>OK{R%BQ:i=Jz(g8=䯣P՞ C9i<Ε Յ8 QJ5"K)zs<=:tv =<3,1- 5$&Y˵TQ7PKwQ߼3&H|28.DfvN҄s -"W븼3n.at˧'!V('sSigk`i2n]I46>[n pP+PlP i1z/֣T/ U| 8; $Rr] ؐʬq<P RGG>zE|a5=fe5_}mH-gz]HR U;SǜTG2>8n/>SLsPJjS M<+Q}ߥ*5Q( q;\WjpNotL b$6BZ;f#A0GO&Dd TrMܣb1J&i}P!^HY"!(O'ceek䴐$\RTS򨨴`Dk)G:tdRrt_!Y_H֮7meK-wx֏zxC~RdT>ױopl OoGWcȘ+yZ&\rڢـ9[8Yz%GQyM_8vI|_R\Zw~1 SI=E>*gЪ;Zfk1f5{?k;EÂD4*G:[ߕv_8’&sdz ȅ0P%'/hsӖl(jY{6S3[I ԕ#Ml?hdA)2lO1.go#d ;ι]mv>hyryr-}?-͟)l(wo{Y7ڈLĉVF}69O9s$)O̯ܼWR woÏ hQ媇"Zo{FA A_`Pgږz%XgrLn'RV@Z7Fcd:˅(XMdJsHAppah&7,3jH]%U,u` `A牖4zMnOJ.79[}wv8m~PSYT* UlXGUk/9֫5?Iz%4er,b MKL/kCx)#{B iiV7ڢ@g:!bI x5*&M-!68uLk'g<Ľ4$ƳL=.قq|]U*M Js@fDB͐K:lO*Mek{44<86#JjBݝBJ֕u#[Bbk-ڒXL) ^a%PKK-dIkuZkвcG?/BˢqN7zE["˥`s, TH]G\h&6`$pȉl;m<8v(Cd(<-!Zy~u)!@YJ5.GW\hX{-[;hڇ4a:8.0ĉhp\1>:J$+5]F:Ƞ5cO^oa M3ѓetDslu>Ȧ^ʩ0)d8ܱp\JE3Z gUd#* S5?ozo*[p6A/e)f"̿R 7|I&Hך*[*q:bo0{\Oo{|?˯߾L_ۛw}gp]M$;"@#x]붺]s {t-QG9~]{A&PAc^q96Oz|%gK.BQ%S`}#g>EkpqV|.`<ejH/'ABA< ӧRt ֞Bv~ds?b@#nGf :Qy$x>)BYJhO*ԀoACe9it[Bh )(m)Na6# X.OYw )4, IvxL1c*(6c`+/oB2Q+S36q@Fw>g؞O뷖Բ&Ne(du*+a]4Z9*m9 Ty|'&e >$sKxSU W9b>J/yMX*QL YV*6 Y(X;Kj1 '6LfQWk~NL>@7ڹ8ڲ>!3Dci}! Vݽn ~Zi"V *e[ 0e%]ҥT hf+HIb 9~2DPޱLkFIKxD=r;Bw94*A ޯ{8b;kEq8 -4B[2{%ɴ%\2D)%r,$ 넜h3YՈ,5r̹)؇ڤ M F)cmY9unuX0b8ej 9|}-sRݫ6AP5L 2hҮVY'܏z_r5J*gNk-H[/$kRHJpTTJus9P=wKzh&=XïoMy$2>nwvX׿>=bJ6~9NŸs)89= ۆuɣ4j.ЅըUW֘l&̶=h-1;Һ*rRxJ|0 !5x3|3TQ$n"y  u֏Q63UApŹ\.{lF:ZIŠ&tAԛ/iiGLj4\r 6`3-t g0ٲ(c .Y5->Ib+?XxvEt4gu__9%~||(DǴZW.6y}KRڦYhEs0 “ Bˏ1`A$䒅:D/P`r\TViׂ$* @woN_sGL>^А\[Tbw5+--O^3Ot%eMF|LrFG8PY$BJr& w^$L g+ ֐'~&+{!"saSPγ ψI`\}ϹAl/$zt b^'ЪFE#&Ô19iQ#g'ꄌ7ևb?jB5HQme7 HKC#+ ^ xU*胶OiaS<^Gg,l|Y\(D6N"#_~|̙䒩>u4~,ܱ OAQiF**?ʯz{: gCw=C aϠQ{T8$|3nVPջkK_Ήtu,A~HCwVs5?Fy@yG֮]ȃdC۷ը55môR*Ŗ*2o,J8-YJjzkگzgX~(+yJK?Oꇯ?~m?V.m:s>m ; =.4Et7"]MZi&'P~].ŻO"O*YR4Dh8W=M 16\hA2lQFkW CM_>7_]Υ4tDk1f^~;~kLx^Zyd$IĸYd@ЫC,J8L96_M&twfBzHxZ4iTznt4M}mcߓOd *mt*h=ߑ+ c WnvBݣbˁB6,\JJBJ1/A.ӢN cfQoGmHzXn ]HW}<2u 1--Oƴ蝹_LGEa]/ix.*P,4C$b%DXhe :CXw̃+wllJh IfQ%tQ::kO/^"'xjpJ̥U2>FSr0ǝmI;=Qy#jDi/Z/=ټ煨'>tqZo0jgѤ-!GP=OFќ*!kip-&"D,6f6X~Oz L>i02IDm\Ah1tqf{;Lnn_!f;г,<+[Y>o|~*;[nu ?PzIS[z\R ?o9-@^Ti-0dZs<ZUj%c)<8g?hF N-৓<Dɜ&9x'SA t=j90N>]acг<$J$56 Z2B 87 ъף&`Q:pM]L*[I/0y9D #1B*i-pj:c Zsjsi^ \$5OH*_$:2"qM JV C>@MV7B_Sɩ[O-WVfTEwY \p๰јH͹kyۤH3ĺ'51ʎ7f}=)]". mMBM +@62Vk~dUaaq, _޶Ua[n.?סqWtq1w@x-8b{cT&$@gZ&!YA'cT<41u[Z=66QCЀF84. p 1exR9;L'xOb jWz@`4An;.缐Q=DnGBm30-u4, J8!%gћdUhAL ma@eHYr0!+tiA$o%>d:2=ʼZst||>z'H+l.mW'Ļl]FlbY栬I9?gZffS2cBF2+=PIr_i%5P7zt\`wkoS[2 -!KN#Zz2Y3joy)q"o>fAYS,v]1(RqՕ[oeR`Ub#S) i޽_3zu_g7:_}׋]m^S&qjB2yNpp?bd`c\}G{@CJa>W~Md'u4ӈL㧦Z ד{-Z?.cs"sjh ԥy[|waپxwr%:yuuM5*Aok W1;qL^8Nƅd;;WPge?{֑B)"X`7ٍ#Yk"X<8qË(G#G b:Uou}pvA;{?"ZVE„^E -ǝk1~C-Y {f_k%'} Ǧiykˮ5 f(}@WŗvDw Ύ~FY|v5˾^~Uv_}z_w9;4.?;YJ!b+jQ+POZ#.o_p 1,YL@l *FEWzP7E㲇TA+A@zcIMeV7 ԅe5x9br )dms!y V}֍N5g2sy~4S78R7J"YU6FguTK+|FTo;ʠҗoy(^ClzRo WHև\dH[V9(ODFT=U*=!:Wc>A^>11&rV#(;Pҍ%f0kټ-ZҝtCһ&ch0㱶X6KccjnXᥱ?hi9yůtʋ?w1_w܌tq1$Mp)3wJ[r=y]L^zV6钖&[} He*;sOR8ÃwRoMTfO{׃Ew0CIo1g%su(kBMSg}0$ítаև^/_ꯗ?[.ipJվΙ2 fl.NӰkeg^%ܠ/f? }B+oL:⨮w=o 3o^}Yn1%UHse'MͪjN,w>yQd(b+KB':ޚy A%ۢ2iN.p'Oqwp^v{uY񱋹:MnSs*ySEi̦gG[*.C2࿶rrZsGث6/-KXGx/HD2XI>A'ngA߾?n?k[Y%*9 \oIJ!9剗.=_rL@SD?4 Ұ^>*>zu`\6Yw?ihw׺mja >Z{>xN,-:-m%W H}-* BR>;McttXEУm>rǣВ>BSW]ᛃVQ~<@h#[mQ #:]P*ldhZM*ф2EB;AO9WUVtR.~Gb[ZVݍ-I*7S\ƹZ4m1j\N޿XК߈/nmYi绥J6oh?Q@Lj1 pT|6Ʈfhnh.DSL-^rJ x3P_˅VGYVPǬJgwn( oR!TR2玮Bj]4t +ރ%vwxfg@+Z'A裃"=LA}:}?Ү>>"UOڋ:fxyF9Yr6 b|jŻQ$"hj*uoD $YۢDwAG:)@ÄZc*Ôp3vYQZZ壝)b{/.Qc `ҖBvڒ=5ڐ]jy;"4#H2](O:elEH9gADgb {GWch:^vNq2Qh?(QLnBWˮRS Eɳǚ[]J3qF)ξiIiZR]"(4`k۪npWO+5+ thPe8ە\|8VG!HTP&׀T߄L 𷒡2UWHPcYd2j2inB AvEm%˲fH57]\?A0nPƴȆ:om@Da(4VCH(,P6{ #ڝ1(QdFtQ<$許t ̈́#Bl `L1R"%8TRp l'T bX_ `I. pqT AP{ T4gB@Q"!\`#);:/ͤ-;M;^TuAElD;TzFy}W@B}M6y-dtJ(46A j"|v/UTtr+ZQCk>84i\Ƞ xQaXvӊ~I'ьY7ICQŘūc8IB1!bE}o2*߹a 樚ص#[Pw!t$MQ #&[o*qi<olJ6!+J$W{HT*#d`!&SQd֣2:/3(Z RI"i Qk2a(qcGMhx |Y%H Vjx$ۑ ox&cQI,TstiC}^jqgE3#HT' O7VEd92`3H!/&sP6DD —9ttIgmG]I@SQA`,C)9iD*ZuIBvp 䁀:t) P+P(vnj5$BjF,[vicPН%ͱk?n,L$&8sXT*L*BȲ$P>fl~)PNX)j(6OvfH'mY4g#MGh tz`"-EiPjj5]ZZz!Hw.oi,e4 %l@M>ߦ%LalmRj6N:jNFW6 'zIѺO34%7aFD5>d8N֞SǵXoCnRM5r7tid"|9:Aff5$2BECt](I WKhI҈5ץ#*w2"aAA(E_-^|rru&8茹 7[Άg߼yĬfrW 9799d 5H.wxuj Zom:`_.8qla ]G) X.0l'+y@Q/W:?9\9:WG\JS \1pW \1pW \1pW \1pW \1pW \1pW \1pW \1pW \1pW \1pW \1pW \1pW \EJhTO*ȧ\'\J+E t+b+b+b+b+b+b+b+b+b+b+b+b!++l]|2+XzJ2p-W47\1pW \1pW \1pW \1pW \1pW \1pW \1pW \1pW \1pW \1pW \1pW \1pW \Y+KKOV< +k`F2pW \1pW \1pW \1pW \1pW \1pW \1pW \1pW \1pW \1pW \1pW \1pW \M \}F0o'ٛ襧(0sor;6WPnc1٧0jow}8!Y21@z0{ݖ}6lg٢:C}>et\T9Y/͗[ofcw%m $H} Ay $ JU #PYC-ΰ:*;[~ߊg213WCՓm \YæP䓙rNzҙgq?>h^?6>M= L-& F|` Ku_ SJӛK{6}J"?a@/OxK8<ȠDe6TK@@rF5)EG " q$#ֆXS69V(htiLHهP.Zx_e@Pq ( ;*Zȉ{6uaB7) P=0<̚-a1Q>G̺9G*?M.9Qh 59gc,c\;nDUMG%uNVO_S.I92 ֱ1|64$AXjʏ~=sim.(tgOX3((dQd4>|<+SvkQwN{MA5̮3׾/ 0뉊B(kB.0ִikC:Ebh~u/=j< SXR.&@(|=~Q'^,#_Z6Qo:>*Yza Yb$ƑRZ IL4"7/#rs_?E864{U -~=}∬vki+mŃ myvgjIR/fDPנ)!y'w)jk҅-lf!l2t.xMpk7Z.ݶ2I4486J:ko+5ZJ9{T H]\ɍsSw࠭66 xvGFԎC⻸Jk`-e_J`KHr~'嬍=oc5ttOo -J皶+ Ѽ͊|݇slZl3_O*k?d Arw[r۴ BR):Lh@잾.2oݭ `q+pŇ,#juXh;^nrߙq]+jr2rgNv9o|N'Y]>M{l8KPX"Dl8)]Lڛ)>.Z;>&gt(#2@IN>:+q@&tY c=LP6MtP%JuEaԵSގKXD>dDG8𸯗 L7!R\tgqtq;xK0G({L~GWm舉 HL0 h꽘r[~E]oJr$%0O7)gZaR VL@%3am*\EY @?]Y`}$Ht)5iokF񖩚'aZX'h?itgz;K/f;1 %0뺝Ս6rR`K ;gy;v6ICꢌTX|zwfr]m ˭6ɋѭ2g|I@RG8)k0b*iӋFAr#bSUbpcPWaXM46,arxf#gXTsC y3:bJ N57GOVqkX[SAR襥@`G4X*Z%Ĥ#Tp`Y [A*zSaÜ(82ҁ*ˁ;(d6,`i+L7Rc z y 061-RkO"y a"5|=[_ 7Տo/ꃗMpf0I]f`>߻.ז6$jI^^a0!'X=1nnʼki2kDBGi' 7. zm{%h{Inu\)zT(,dss}=4Y4>YFZrD{8E^d?zS7o^~jDK]He>dFɠJ]Qh6LUTN7,g7!Ϳ^_z7ҟ_\7\xu//_π.!Q=V]u -Z.g]-x~%rz_rv  q? дRagu妍F&#I꜊֨Z&߂Y*xi_P,JCBkyYwSZXubգr+r't+X; ,6rlks~5P MkTMlV0f9@s={G۴2@ehA-tHl݊?,pMRBaakIMmi!GCQ$ vF 'lJa^iGjd&)Oc.*5XJv:n3vr]$ ]֜/]RȽn41-0*;ɰj"qnA"ĿnQ~ avnGtDHI;GG|._}?}+H9sB 6<-ju?}Oт-`9"}ϱt -Qբ,ё l`"Dey rͨϱ(;E2(:o;cƌL"^ˈiDk45DŽDt ;N4#8@x!)ũ.p0F9EQ 1Z-Zxs]K@/nسo.D؞G^J6[ϴ4NΦؠҝu3Yڜ ΊLjb|nt8@T?ZqcǜYk t0zwkyl/fh.jY;t Ť(Ss>a.GyZ(R@ JB1VpµW|5nz ҙ^([TˑpuP0Gygnlun%Pn>m*.V]ZLwq~9L<&^-x8pR9fr'nDNFh[E5߬ooz3A+l7 VY-n%֡MG ?dlo:V {bl9]?ooos8hΎ+,=}5+nmݍWj[xs;? [<s.b˯h,^^p~dʬeԬն]Ɩ|L/T? :B=zJ6 G\9zڞӦu@WYAgoNw=yr9[LvcB(r;>Xw@5xArk0($"5X@蕦!rY0!5creqc EĖPƽuH ÌМ:NR/,^_[r1`'[qRM =! SCr0iqUe*wg']VULc'79_2`E$WTzYsΥFP9d »X"WVPGt zÂZAx Q;ezOcd:R tF,B0I# be}01SF 8<Hw :;N@+:K6فs|_Я*),B@(UIZr |XeHFf%a> aEp[(uBvl i-H>E9ɈH,eqBHr 8M(rn%Ew9lVJH,`kH:F g7R ПIAӒ%ȅ"}`G68B̏1_@L&^"MNi{B#'B#'ϯ:U}3PZU|7`I~]-΋rF1v9#~?TQ0{)Q'o{)"kiYIp {ߕТ|CJUuL8@")bI y>Yb:bhB%Tz`2DO%ԒN>$%0&][o#r+6Y %9C<,vFllbK%dO/VT7UR l$pV X[Wv#e]\+VX=NI|Ck%~o|ݳ:nKjq6Ao5Kw-?'>_ه*E 6o[!w)"uC4k泿J~*znϱ|G9 t!KHxdQfTHS2dCz,ot(v:A[.-Pp19D)2.H \,+Q!⿍TNn #0f΃1c˜gй#5i [ dwc܃Rњ‘1 =)]h)f%[j!8JXa,bJC#iǒGM栃*e-zD]McOY&-c ^sGb$$ŁU%XKL'G ,6AN-T n{fp1 ;X 18-SF3,-/Ŀb͆ ?VJZ޴0e2R%t&!-pq[77jܭ2k۳>#u4j]G^?CNLƷ%~1XX'eUb\Ѥܜq,7"`yt;}=f8oC3T=zrwJ_{O|-Vw71u-ۖe/.:\,/tQfؼ%DFB>AE"V9I6@-3Ȃe_}`JP*JL2hjasoAl^Uf6=UN)Jx;9!Ls.uHU_s$#=ɥݳY}R_qyw:gN@Ud-}%RHZgPtzjxRa=Q:49^$z%bRIAYyd&ɫ h׳SqBFCjcġm&RT[]\W}R /j z7 09 y@Dohs)!,ITrҍX+cQƿI,R9=XeȐڻ%*'<*Ld9'-wD'q#)AtƽrƷ5Rb;k}Y|_c0Oœm+@DbB W 0S`^\#0d[k@zЩJR1 vjɘ5exb ,' RJ<^3g!frT;SChS1/AAqIW3iVN[¬_O&<7]Wa{q1a=%ò4'9p1rۮ.~=O?-IϫMN>x\8 Zɲ9r# xn"!@`E觍ug3b7f}=)]">.g6ZA&ՆqjXXmf싅fW#*mqUq <.Flh4]oZE*}d*8cL Cr+xQCdddɬ6( AHQ`S Fm&mǜB1eՆd$橠vq*mc:&EouI 9l&d":gB̕Ӡfm,4Y "e5X4"  T~05H:Ն;Q9+x2Xk~싈2"D4w$ZT8K%ْA[ýGQR:;ief>XP {3V6NZI\U" S$íL.FRBY X*#b|Q:):͒}qWE7​SsLl ,y!ȣ9{鬖L\a$5 d\<. V~?<<mILۭ[bDZQtA; яG~t9W?9WS6dr2n/՗X! &$-"t{2kn뒂6Xx LCpΑ +8XAj4N= ϊieI )9mv`PW "fQE%Y, HВ&s%8I̭̫ g0YNg/_O(qw⎴ro.$}vCJze6;%DOJKV\Hy0Qx͕GfmHVf hz GJ-M)[^Ղ: nW8$Js$Bv,9V2Y=jn7`6,rOg־b0uŠ,,-sJQvᇞ[ne!X QLɴe M_YNw_:ee:)eOwϬ簗aMB xF, $"4ZruzɩpKd#(RJK6z@eL9y4$65 5&JK0 `TL"[,)f]{uB;,Q"4EMك`~޴-TͻsZߚv\*g9ϛ_0LoS3NfcÙ^Y4QJӆ/ͽQ_~^kݻݧ{KM]ˑiWI~ ˎa5}*Ŏ/HN2ò%Le_j__%ݰnJO>ܽq?n%M{qTB2^y%ofTӊł]y[?uF5Ϛ3bF5f tKv"ټdz"N7^ '7{-&4ZTu&3c"g3jt@b ԥT]io\Ǖ+ ~T aĞDc$ VVؤ" }}E& |}.TթRnÛݰ+^a a#7a󱛞h"m/On~S%ke@r#)ByV$~V0-4|8\^W8_W(bilUD*L--]h9h3pj`g_@CH6y? M,9,4"cS%Z4ouXXvS(*7N)u=R\1]S"˿ϖǯA-}v5+WO=kJhչJEGc `ѸD@JXoRwōP۠2&UQ3:XLN!m6$3/֪QԩLf=>ҏf z&0XdU9i{S-֦c.kƯT:ʠӗx(^oRlfRo WHև\dH[V9(ODG *Rm ejp壊1ǘf[HWKehaJwB.dw}c9~;1}0X#־⥱ڢ[Fxilʥ1uϥ1u ˋbZ<;x-Y͋ۑ.錎z8{V[F,^U٫1߼1cwm.~v=@7 Sb`+v\&37oz9um&xsX3jX|a m0ysjl=yw?kܤ1>CREqY%v֞xYn\(SC.}kNQ"h 9Fǃ 'n3lr^[HNIi1w>Dr}jVjG(T`jc ޵ƔΤY]IҲj\>tu*wqtF'%Cx2Y>Nu+;ք"M&_q}=]O킾IHr.^!Vۓh02#ylۭcz$V#jCp9㥿?eprtyzԿ?v?k]nJ~Vݯbt $*]9MohwpڄxzO^- 4[ާxgWm|h~_ N6mv>gA89} 䬫I(cݤ3*@}]U%]bg˫ T8/4<̗ttX1VFvm6oWU!ǿ8/QH~w7NvD]nl՚YzW܌Mn ㆪ+FRQE\ \Yy{|\-#q8 a^jAs/U1]w W/^n~[R&Pʆ$FM6HqZ.S;Y.$LdZvNu\/oÇI  kfy{}Dq?Oׂ6KtLEkNж\rTHު4݋Ryc-h&ZCd`ܔNqOҍ֢kQu8=<ˬG|wkKJ;-ݒMU6,@RymASp- wO×2:`0v7cv[\D\T5Ԙbh!SW=Lã%B}~[it1Y !smwH$+Qޤ*CdC #ԎD&xB. -h'nK>fg+Z'AF"{pG &t]=I}d|HEb5Zڋ:fxyF9Yr6 b|jQ$EU^܍S5Ժ4wgDC))IV:ѝCD\B&}d S)[jivJcؚ}'餞.]FT,CD ɞ5))idw"R5V&Y)dW(! QS׆RDxR7q SNiE +1ؽ*؀d@>ѭ̡]K68U& 'u$jɭZ% qc- `I. pqDVZ hަV J8}$kG b Lڳ\zZëIQ@IARVKWgT0-([jkIk (D$i4(VS Per>@g C a X+=sNƁ xfp;`ukktH͘uQ401ƄULl$%rBDE}o3*߹a򒭊UIՂG{{x2AAI$6i%‘Fi:Ҭd :JrJ%V 0 CMZe&5:/3(|" H&jZ5 @6!+*]7xѴy{KH'*@Pǣ@܎m}K V'ES "|qgE3#Ў/Z/kI|ǍU.h\7Ajx1) 7DD.c`rYwk)1 hcJI%h v% Aر-PxR0 (-f1c &KpZI΃pPJ"4)Y c~XTי4ILp҈`*L BȲ$@3A6 ?Ft26Љ6 4n1V =li`F+@$޲[PRS[ y/GVmuЍZF7)mU(am싶V|ZR0 ZUK F㠝nlxޯ|aNiۜ눙$KԭGwAWwZf= ]Zp[ nLS$J(u6tk*ZSRu:- %'ޮ `L L9:IѰ[N 3bO*Vh A xYh˳sE5* t#"B[uzҠD+Y.T@=ʀ Bj0FQ-B>A =-ZiaKP>]Mw؊PD4cҀ'7(3`Q' +U 2P(EcDMnƂGtvy\:`D`pQVc- ̩m X{Gji&z@IQ{/A1Y7/Q.gZXBv Zx:-Q秝 v)Չl z;k*>@`HqcX©:lti =!W&EBbX?H(Ar|d8Nr)JXO{{ݤ>jE#WnD|9\PHhw#!:.P,=uQ@Ւ-0 Uyۈ'.dޙGF'%Հ_o/j/n^,.v1Wsr6\ ;Lj^=&1Pd(ju̅ao6c Gqp̃>CUSRh)Ӛ'#+F%PB*V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+^%PDyRwE 't'PQ'tN Tw}J +x+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@_S@dOF DFT@V ؕ@d@ʻoc%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V))tD2z:J ̝@dm^ QnNR DwV@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X J V@b%+X JH x@֫g?VSjn_o V}? ۍK r\6_Zy3_..K[ܦe38{iUgI۴ Ayi:6 Ju}^0%pN/Oi1:a5ub}Fc1޵6q#ۿš/7dD⇬n,gj.<-:ȐlW*60|HJpDe qYg8Ӎh @ F*L^m}Ό{~N&ۢ/SlhB-={0O!s\ri\Č<0BƙR~mr[k$M\ δ ,1:XU۷B|>зB3Cq A橆>lBȲPPL+d%!JchCT^fsB!9aLn"+a1Ƽ{M:MDÎfB:=::&ʛNe`/j@—I 'U"U/ZӀSAfdmN >yhQTlͧzxeI %xx7dgs4\v:Q {U0芝,qMdB"M,E<,@x8 (wz"aݔ9x<}wzQq%p6oM/h4UnD EQ3(K\2)&($BpKN뇔9RF}}Н j"&,fa dXD > y7NJ}30;䮳/i^dJm9܎0{Ͷˮs^-2V 9;]^j7#Ne5^j2] ۓ"olv 8w̟wLe7ͶSXw5KtZ ΢'SUg.q3뤕Pi+WW^JKdIhjSꘊn\ Zc0S{ԢwSI14~I^OF1Wے`RrSe)+\1c;mLT1RB+*(aTZӺ%w< SR&ԚI\фU]ˑ}=`V .fcV +aZ2FpQLZ1:۴*5 vJEa&) JX+"!m v9c!Iw=U kf9d\ܽ#+ٸ)Y~܄ xRlB~&E m#F0{>ҙ~>pS"Ѐ=/pSq b tfLmU+i6xY菳$h`?`?{^@8kzہ)eO;Ϫ%w<Q d] fC}6VJ$g0mºmK>J "BζjTy[!m)UhZ#C}2&=/ 1,0rϺ%7Mq094u\ 4m<%Is)cY>zk!Zr H.Ub#`$ΏV-L"n-%ELM ĩ{yn(o8*F# ^ $8!vw=g N)CG2 e}f  Y[xt߀o-V{6#EbgEJqSLKU&,K):IiigS2">ےZK4ׅAE6Yiy.ڈFҧw2Mff?!V|Upw ecȔдfa^,|`ryR^M}QW؜[*0u1SIHG$/rip~bSX"+f* iT^]Ua=ĞeT8/˞ߔ0G?5OVP2>~{ 3پdT7Cs!4EX#cWW;ٳf#ʇpg ac<(;BXjl 2е;h “!C8LC+`.>t?55΋&3q9pB\yxJ\v=DTZ=槁(P`E0%^ЋvQJ}U~XڮxeR50oʏf#p_f ꧣ 14 c$.I\Mb[ ADY1KVQN9ɕdD<<, T:+yR9hsnџǽ*^ܔuEΫәbd hm0wE6Im~VO1ؗӟióiuİd NrX,C&7k|ۥѨfS#"=ONprοrϓ7b1;R,rw7^/'#{o@m~l8dχ7Eoø?GkxU1-;&VB+}_Ai=?"P\'x~wRL7ee.Ǖ^R A u eLG:!KpHsYqYJAci^ Wް7jŏ:9iP?㔿A֕^JnS~8 \r/)8C"P~)?kx6n#M)e7ŰB]&ĕ2$tE/R7%Ru.2qӠ~n#HH!+tR:sx>rֆ2|82LSi0RMSs_i7:g ݺ1#qerXq8w4e&Lo ٽBNs;'excw@) !kqX8Ex-d ! N?'erNÝY̦O:pRLn92 @/R:i ]2 Np .2NCd9+ӵp'DC;xh]g!N5:(?ئ3g-&tf.Zyd~9ɴM&~ J<#xB2D˄楦ZY[⢃zzZRzj MP}!˅ڶcVo(E<[2ΔJ. UZ0HJZ&IR^rhcAR<ܦ yf `n-kRENe6-Fjas$3T\a)s3B*NR&iV$DQNF 2u^$un.ٖԒ< yFl?&jϳ]䎟._5޶Et ,43o2X{T;a! LgaM1€n]U'8T^S9,T@ٛsplNRhp7Gծ+9ꦁ:X5 xIn(pNWɸYG(Vi;xuPP~'$̮vP~MͩJ1Cbdh ް'dQݦNp9lDzUU+XvVLRG9T Y`ŦUWm kΰa 2!͋@uα7!x~"w^(9ܤoqnqUV`:hBwj6v]+r9 Tʒ% EA2 B #J DKLKƨ@!1Je3r\kJIw/pF4!t:ٟȸf6LφAwOǜ08^`n&d^\g_.$Ol;-2s_bzkzݽ|,{m<1,*Ʉ\qFNwD@{o̓hW #/K1ֽoI]w:/lh]zU?;վ3Jfh\_FMٖ`D8U|T !zyv3-RkC7"8ޞ[16ב}o#-g{nȖ4q +w4f+!%2ᦗD)hg{ݔJ?"ٖL{'U@<o&Uա`KLB08P]/<8fk lk&pKJYHy3DaDLQRRTA,Qxky:ɺ҆NjpݎZYCd5fNDԗ5t&m] mP2_' x|U-YyC3m*dUi2ohhh熜c?M7eܔ񣴢kWdX|UȈM)%iσ/NY]4@OrTdMB$мɍ˭2GE+8ߊц$&C0+?9#2k#$=[VesKN3 7e(vgСXs2z]8K鼝 ˇ} X_aou;j 8Ok?Ί- p ̎bD'&%[%Lz^L hĊ@'VqpV#.)[w Q$+w @ ix[*_&ʹo擛 ]t9O}LZ/d4|Fy6ݽ2kh? [ofEU+q:آ,Ubd2^[^YFD&?!hOB|M3HdcU$N%Q()iQ˲%1;M1n0DXhm&{M& J($&Nfinzm_ 2_߃޽x_wZe**\MoK)]z>"MR-t"@qr34c/M2MSE&,GBgL9h" ikWCcEYǬ ~2¹bq6°Ц1cŭhvBY ]DGtՆ33Kkp^26Su)gzSI_)HQ=[fJje3`_?`LbH%qW%2f/KqL4}P2ժ gzrLԴp8Xm]З8..6_PdȘo.:7BՇmK#S_;Njg$}-8ȉr9 K3(=l # HdOK M1 F3 b(_>ٿrNg'*$ӭ\)T9Ѡ $hs:|]ȃ%,뛼=PU@ ih4 'd:sʁ2$DzgRb@" "强ҹ(5`T 3/!u ȓH{o2TKY@'iB,[x%|{%#:yɇKDv0! =?gk aޡ90f:N"e<Q>s8v&2{rq*)w_<]Ũ [䀮)k"@<Ė=i JZξao;3I^ q7xւ ihMz a9j864,՗.@bp(p4q1B /N LvnZ,(X9OG`bˊ(fqmUODb9 \`-"k*\s9orrFU\@₏}N]HP@~3'D]zrVE;_ Bid@ )ڃID]ivjS,uI7ޔm9M9Ⱥi XRưӾw1sHc$X.b_h7yj)^>)8OIxNsNC0HPr=! `]`X>lǵ HyS zÑ:4U9*j-B1꺪jx>#jz[J/b80{u/kn Eh?&٩"_;+cTtd"UOύgӹjld,Yt;4OOeBǁdɗ@NRex[uh @RJ(?W'mUJX45RBi&s_(ѻ[8S@"'aJH#5Zś4h_"x~M 5Ļ1J|%I 0ժg|4=l1x<=Ҩ-y,xt}lc-8n#A}0ȍ05 1a̚58~E,2D-̗>axKiSlҹQ*eN73f|*{?Fh>]&H itB?[! B\C nØ%$P䓉J*Ip&0~~N瘵pWA5Z@w ‰JgT)>RjEI".g`CFFCu'mwV|u"R*&dl21[ hد\(9E !:|]^:f w˩Eb oR2ρ+4/2CJCS^OQaR Ɇ'4a,n쿿Ku>]̇?_^vv6Q6tIj c2|~)߆Ӈ|Z"S(d^"J 4#)F k??Vfj&[a!g*ΡPsa \s2eD:3j۬{X xpn^965Nf)YKix58%F&&m%kh,vzWyPʇs>NK4Kvw܎2hQ 1 D+&z(u %Q@^ Ҍ)!sg *J$"iNlY Y %,UcMH+s3L TҔaO`' ] 2ܢ 4m&2O4T*B#c'(WDQqm\CfIBk/.oT}]ROg_ݞ@1Es"n(" J3UicBsW@6?m^]#Z:X^-+x,f,/&$SpB{l4LMvV4vGȷgWao X"OBtDwA=HV^#Ͽ;W 'Րl5pn׭bKU9 1YAs;'bNo d7[W(eH 7 :k}G9%cf| p'9р3a.Q&x iNѕYQ$E԰#']+(eD-,xyx/솣(}w^]GOS/,'YT xhp#M6GvFth!gZJùjul9 VmE݀UC>$T7fcGxvty94n77y==er>bp"MSt`Sт.8T &:%߰ƽY@bMu6 % [w&$BwGz_D+ٰFG&io.+ˡa{2!~`V`BG~@cW'BWI>3JfXun&,fdjThdHl1)ƃ H5% ķY]6wm9c4|T7W=_gvj Lo JZz](]/ޱ7*7Ll$عnsuuo-6@,?Mxa{qKoWYEὄwǰ:!t^ t^gku|S~Hu%aE#rd GI:*PY$˗y1H*ƳϪ.dڴ]LhpoDm ۹y{n/k]sl1{]ZЮZZع5=.F 3F{otmyB ţl RaԔcؾxrQBs]P|iaFk ۅ+ lke{:ےHrZBF{~փˤ@;>AK|ā7gHIĪ7ZfQ.DOJ`kth6J'oŲYJsJ>)ra@*X.4|$OS4jF9 0e %x_PP%ϥ2#|Bݶ@RǸ.υ8%@IVFFr8*\Hx' U"]yRRLQ_6Uє>n w!Bn\x*.JM;)Ο^+q} ܍]) S@ YJ3#`.M /)Z X-1 y%؅^xq,ѴF:\qcHzַl\q꽊g \\&]6 뎓&m02$ƛyII.XW%yjX)L ) ;?Xc cs| ;޹|'/5T [o'OC?%.0"oXG V>,@{Va{rۚxe< u[EwD0 9qn4>(h^YӰFG2z1 *@yLBɩku`WhVaýSYˑ8 %"հF_ [qڤ@6uha"2άj|:*.MIݎn_jL|Y Dmσ>@gKްFǓ)WTc~ȅN3:Tfv Q>n3_E`R ! >#Ͽ;  :r~n]Co'g8VF]CkW)+%|5~{*,ָ˂ 8UeṖvݭ'J1|À zwE55r>șVie %x}XhJJr'ԩ_~i}yj9>c}P74w/sMb-?*1s^y MqodߧsCy.%-4}sO4v5`PJӃрq3 '["U;}|0La"dOf,RO澧hVm͵_&<Ĩ^:48n2WIu7HTB]%A4@m`OJL@jp0F9'7B4[ABCjfLJ*(оc.*8m{XaE mb҈߫1s K޹AV׼r>2Mt{tA/lTtZEN<P9r(,昄}6tEڵVINH+w },pjr=ioHrEX a`lf&AlddòHϋaS)5$e`GWUWu]}^}KPh~gxgG"oAt~" ѭ'ߌLqJ~材s=$OL 0~yC#bb(Ko *NJmR*nD6Oq@alNRӖIJ!3t̔fɺo!{E |nBd`cX8~ ^ԓ!!XbY|v\H|lqDHYiBXX!r@̣dVY"qkg$0fVCIxKEaDcz"[T(|d)ѫleRhv #O?y塵i=L 6"9lKm:RRZ .Ba'a[h4 f6{^Y%l¤//:(V3`> fW!8kai޶kzXAv8 id0v]kUtsΕGha8J7<UJjSC.i1DIzU{JlFq|:-,f@i*U΁ FaaFMoyáh2 bƲU]Ee%:D΋& r)ٟTֿ0_S.>(䠚kݱ?$}V9a Y[U=Ir/_ɘ= .̋xkb5"Ո+ D%"G{M""x&2%z7T72Jg4>$i@_|6B[/tk)q'Jg)9}V{954 936=T<87ꎑ}*Ih׻W WƺؕhW {//ѸzqgJ*b@C;TKA˸U> Wt2gT,cT;GHbpP!UÄ39c%AmVDhzF[:[ΫCОo:zttsqkM3::J%:h]EƖ 4T^ba]rwJM:>_Q=Q+*Lߍ4ӂنx;t 6Ұޠ|Lto{FŸPd[Z&c#) w.Triʳue\4D{yZ =_0xrǃ gXw^ѬTj IM=t(>l@4E*Hп,CsE;p􈾑d\! G: /H&e'c=W6$wg Y6-8wo.ICl8|ٟ0}[4@`X|`Rݲ~,Yx+ ?fw`-FŷA4b^՟77E1ϯm2 0ɗ \KΙ#E4GI_U7Kx[6J'tŐ n]/TRY/tT 0#v:?w;9-e8yB,4ܹDutdRi8=kJRU#ΰAbdZˏl&<Io,/du֚idh}{t ä*WlZL,&=ȱOד!;h>< ?9FnRw}}A0i۩r3i8CI%k.ŽzP& TPXwWoJI)ڐx2#qنNsC`c-]xͨ({1ɺ[՗,LGElz¥4RNmf<?̬E敔 QQm:CX9 aM w[ɍ2(% !tWTKշ؛\ιKjXz+1%t=0ӢBVF!K ̘Ovh8s3N` mL7٘ϴL׆vq$@u_cޞ2B+igznC7-͉ocÉ>W3N創rMkZތ%)TX & ny.~?6 "4$(p i8(UiB(M8Dk{Ok)3-:nք !STfYSڴr㮸Lc1I"j`'#İr Ol26k50>"h;U6`Aߏj03l8?ü ǵ]+yҟ{^p0"c[93p%Jdbt$a#@nYFƆdML1/ʆ`˳48c ]k;y=Rü")]qq|ϢĺTePg0I$<!rըk#&6)1*!}jBJ+ f 7dVrk)4Dd3%eTۧ"wt0Vug4^Rp͝=`T"QjjQO&|;k<ZV!30AK^Gjly0F]Z Rsƌ<:8Cb1mU 36-|D@+S# /72&*q~*g m pޔh9rDjAk }#;~x9)O@~DXOo )~uѯ<ՕGˣIiBf#++XG T j MiT"%|7G/#rq=ye͆g# "* MQKh \,a@>?]ܵԬ~ٝMQ y^0C8֤4+uu寮ՕeDD *-(Us TiIXT :g)xdt-F x( ~s,ݰmp'pO7*^M.=M8bjezN.Fٷ_/E=JꧦJ"$8M׍-P4$b',aGZ$%9@-gW}Xucsu) k+DtYK?&M)Rf$Y3U{EGsSTɮ] WGqm=V=٘8EfW4fNJ$F4Dꀼ_H5] t a#ٚr1w l4haP8D)w5jՑ)R$MHB2~ ƫ;|َy^5l3nQf8P5R#%dTI.\z]虘Iuy.N"ƫ[Κ/\A!ؔmj[5 Q?gIp#Kq+d >`}x0~Oec\5\bdfcNp,a*-AW+!5f ;P/<y|0 M)2$8U(>,dWRZH0QSɺXW8n{=y(gDqW32eq@ ^Ԡn^o@X`D9(l8x^\ԧkc*EB!؍b>N3+)L#i~,FYc`_ԓ)_p%oZ(0.mocW Iw fD9JmTXV#LGK8ZKU[v_8Te*hI\˗)J!"cb4>S\JMh0b(C b`#Xx$ʸZd)+}@W{KxWM |mnEe|pO#{UrǗXGӹ&Z~ө./ Ŕĭ("&CG`wMqw${?%%l+Ȟ7| 0#bHUpI[,(5jPO@)2O2xy0>@2UD\s =s'ΩW=Ƨ(xJ=-w"rhǯ*7cyž*ltU!z2Mnoھ/O3:%D^[B$OC/$g;Ђ#v, 1.2,RV?;4&Yd֚Xr?}XWͱ c0'='j9#+{YcR§B5h!XpJ $U(5+Mq^x ^p݃J߼ baxc7G|8pA: 00V뽘^ٱ[\VǤ@igԫk~kKWw" yU M a1?6 'x l] ;h\-1_,̴<NsxU :&Ș_csn"pjP  ϻl4}~r_/'{%l m֊n`ou9Iݬ;w~1rtT8zD2Ƹ׆%V?a3_Lu!!؇i_V˿gWYv־+lD۫k^pD&7tFEf~Zwz=O7^tp0h}0w`E]p?_E(XS^fU`&8>5m>oP4* lU";%o^ OtSLg W4PbqB/W{&l݂DlRI4O絉cG)*Q(%5tsAOqcՑH=ތ%f@;}uT~ Lv Θ̼(<ӅV"ф0ﻹ{6>9?6%Y`dxfd O2va<)´ME)VӠ% ŷ\`pPc1qB[T4zƥrl穧r!+0RjOi|sw7^n u;zgOr+Z>gT3\lbڂe6#e/-@XZmlXϐnIx|; _V'I72CVSO74PTٚ'$鳤kPIBPTaT)%<EFq٭j!"M KNHqT|&#դdMݥF~BTm٭yU nQD0}9E;55r半QTwb ʗ.FdW2ffeS'q5DU<協?8aU;|B}R*G7mS_~Θ )kmEdv/ æ{6Afl,^lٺYT^)Z"ȢdL-!˹~g1{+d 1m)AL+ZquL>fBHޜbW&juM~>s>_~헟F2N k[xPrV\UV B?e2R{8=Kp/C4+lewU|3Jj_~H?zU߽|GrB$ %Ɣ[l4Qp?4`IܩI)،MUD\gP.ui&0K$d±6`$ͽͯk|>~Pe>;J>~[>;oQGYe6ބ>W7qXK(-Djәa$ * elXFx}9cGg"a42P#밑&6&Zb#P&8ԆAU88y8׹Îz'ou fvPe*!I1⑍P pHIQcdU9Q=}WO9PP0cRt2 Ps&ulS+"\/IHE6U +JqI+6UM6PVz1>~癀Oď:W <;5+5oTb1yµYjrW0 o_|G^*detIn293.̉P3@&)`XanJ Bf^2n\U j '5sUrǨ9MD"4v. DY,8BFAB'RHZ&w7I# Ežh g!nd<3#h6"NJ@M,X9S U5Kz1v=mfyqOzH(X !y2eNgu}` 4H 5ab4-ѣ"sBNH 4! l204,c>YMVG^x3UF8 JMMqAx,1Qز ^K0ʳ|bw/ jg sDVȏ#1o\su:`L F\i2f;qJ)-c54JL\ v1ウI,'lpncB<5&ڶIGzvozn2>&4p *t] PP.UIW5뽩pyv2 X7M76@۾&!q}8)x˼rNx2$-M&h[#a0jdx+d\vvFj*U+ <݊a^)2./e\bq(@j9tN.o TOܯeJs8q~N^60@B&kgԲwg|sZ +U!ǡ.E(>NjbƤT --(l2]Αm$ƻ>xL6O8,?t1GV]8h JKSӖM%* P~__1&,Aa-eѧ0n J>-U258s_'tynIJJeY!]zfOαu;?t|9l470krX<bxHq p֡OQ =禾+$%"I@e\4oYu2c1ŴeM&6@ȹ Grzf+y&q^4z e,8u/F܋ݰcrhEBNrH+F$2L&)R}&cMjPVd̩ݥUx~9ʗK3-#yU)ZKGL| WZ(\ZveL %\J)sS( d^KK0댒&QWE9(`x<0^:OД'Cb"Hb0Ac3jA\KQ>^]8UǷ]XT!ۮR`%T GWKb5_S`=_ککE,jRX{_u۳/YMb->qP(vPRu:£0 r= am"! iMs jL43(e&y+Kֈ$,mCXgcWb8GJm~FI!rx `qw>9{\bR'a36mHVt&eaZ5@«Y X%uwu,dvqGXJ;P'$dG82T?ǫР+}Rb.*|r]u~π8J2DIYVIX%*&CWrps_LqZ.m\e~1ҵNQZ!ҴYN<ڊ^Dm>1 ?~UΘ 3C'R;zo"delL)6 _QvU'͏q2.2좴 ʛtVC1T蛿CPU *g(+b'˪HG͸9qPҞ=9GAّn"V*MVU#Au5wWNa!͗}Q4DzFs|A1NTZV'TwONj12<01%ƠHkxLIvzQ79-&,} 骘X ;[En>O7Fvu#4Sȵv{ KaC Cab 8Mk̩2}JdjB-C4Q=8;z\WbKm>J%C0lmfFu!‰PD2\O YPxFh,Rv/idbX +3}!@jAPZG/ǵ>=x?qBS,~:Gn |#)ʺF˻tB\x}j-6E*;JL̈́E%V|K 07EsGr*Gf3t&@Qxq_NlV'Suk~6cC8WQ[lHRH^}]% ]\St.{U nk}* Xs35$إKv'|Fk R!:v{zBSmBJ4f堞E H`!Ax $0Ohd:*fmƽ* E?:o%Ml:W,qkEvO#هrkCl]j/Y n>e*Ie62|(&Jƿ0z @Opvء,>e/8LmC>MYVDNE)\%`BF,Bf>xW~pWdꐤfPBX2K~hE:JνV?njx?|2z^ֆLSnk-xUȸ6C:KĖBn[j+S!%MUѰ=/g0F}7n(#q`f:[FH̫|BF .Ծ&L93tY6ӑx +\;T!Dp ^&XE1Q3bR&f?ױzvve]>C#Vmtּv_{WC'U}4e-^1t<䲆UW!'%zdAddt*cu0`u! xBCOIRRₘ)2}} a͗D )MD:tREEk{qHQ<E-%=9=pR]lS^>1ZWB&#`ם7boAn *Zk[Iq~]"`)4qd͐bޤ;>4{\ϥaBM:J&;p$msj \$Rc2xY{YʕDֽ3%&wVi> C(=F+FZ6nbHg:*uNdB*Aٓ5yh Sw:1QKBFڞbR XO hz[ldS7ç_jɮEY8ғDjxM׸eT3sgZлJ>SHE>d6{Ր]1G>IM @5vN7"|wbU'Jջ, E>–SZ}lz4d#)60H n#LD9qSJ>^{|TFv^֙%% J +UL2b5p \.@$;|}GYy_Oas3<f.-H>O?}\ΗGVhu8{~^[6}^^e}Oϣt:SYMjm5Gr :h1:Ұ*Vn4V7h'h6g]n͠yO%TtEt+_հq@Ici(+BFEH1 HRMnb&"{owO5j}8+dACQ:$ZEww :T'DŽI|z.x /|$1KLjG6BQ$8K#%UFٗx-Ax.Hv}fHVj6mcHi :n+f g͐(?% U0460?ˇQt5u9̉qR|_XUhÆ,jRZ _ܕ!( C6RcOL.Cn1B+$Lze݃"}3d|?n:""T.d/[j}T0ȿXY^"Jn X"-.Lnuq|M/͡6Q]>/z{ i=uՙ~,], >ǿW0{6g5/ ~Y"5뺛ѧ^a 6LO'аDծrdyĐ6Y&|t{)aR4Ɋfmƽ*lY)^w!#7Q|;qⰁvMU^Ѝh$;nN2YjS>J{N7"Tf䱥(3#:9< M[x@\}p:^^Ik{^rv(O*dfcn-'1jА+uB%Q1C(u >,]gZtVa&2z Gu~(ɟZwoq)[C5 VqW62H7Ъ sV6P`U)<_ځ 4m7H"GNֻ eGAaڇbH{ fd-DߩWĝˇфWq6GI9%zJnc`z 7w0L2oeΐ+=ݼmh-amΠU@[y<Aތvbȑ< C9xٗ"h:jxݚ]SJ7ήgIZbbX$ P7CYX,-1Z$&Rs` ~ց\XJZ:$Ƴ^"<+sX􆆒Z11H~NAbHVkAű q 1Hk"6ei±-hߤɷI)$1VJ{6$+"3 a1h 0"Oݲ,RNj%Crd[mHEVdFd=ڕ}@oŮ:TDi =uDjdFˉMY0ݥY&¶)AVJ/%zFl!跼_hx+N|"__N1ֽrul!)PcMNXm53#Rpы-0}L|q2kWu-^|/;h@sy;2Y[F{ygv 7[wU]CE ,a%P;QQ[`<|KDFth1}Ҳ"3J~9I*&ҧِȜR`l_*1R֠z϶[<+9E}sX FPh$~#PbW>0(`M럯C޳4}t(݌jàd#A-x'&Na}gx; sO )!"KNIR#0ۛuk-u½?>>}>|y+U<y[`J"o݅y@1fke)D"+dvIt sвiy,Yw SNwP!/ 0\#en?}իW7שiAmN&Kp,I|XYhzQؔ~5 Շ7aQB垷idn"Z8rGG+c$ӵ.As8GX^\9KԹtY,t-ac(MV{@ѓl 3xGIIJN'=# rXO]c H)#?Pt||EC6qEAli6qZ1ilibS#=1cZ6{- p{":þkA14NɩiHy;z@?,҂[V_ss :x\d26g+* ghÀ0Vf5GXV#od? ws`W1$=Co'Ԧw$NᄢQJT"<%nw L1x*e*wzkL{2w<`{"JʩcʳӋ!)}>mݵ@*ByR/kd+O ydžjij1 qF~<:$:(TS}<+= O#-0^JSDInj)ۏv>X-0N]~%5SA~v>j4Y]4q_d `0A[#na#,}lI>Z}m!LOAAԙUIZ(Pc~~_~WMRŒ-,ٷ$ADLX,o@ke\@1Ic1.:h+[5+0EU(kMڤʏ+Oe3';-ZD\t#.wH((Hv:\xK7|xZ=7w^CW;|R?yG#Q(t=%)x0&]`Pk[rl3t1&G)R4dT-QJJ~ 5$6JmڳKqqy[` Ǧ3]OctO- PY|GX=84kZ@S8;uװ-0'VlݎB2#7 )]`Sa9U[Ӵ1RnnYF>)BPJǘ1)Us1F#q [.ŬJYD2Bi Zɤ움oF[.q;|kr#$Xt(:'f |b6N+$˶,x|Dɰd.HT%Mwz0: JgMXecc_)*Y[K H˼*f+"6%ső8<8fa_2f6;h13fn)cĪ tƫ0OB跅XO.*ryٻ3%3?Mׯ~;Ahq?wu3sk`?ϯfl?ɟ起p>??_pj^gq}ɚ?-|]ͬx/bdmkf 3juC|^~[ls g9} H~|/a$qSo]evs%sQIdSUL-_WE0Hak2֑)%YهBrJYvu5aeBlR9 JٱPqhE_&ZABIdV̓b%b,N*Z&Yt`FV0r2$ :8H3=JꩌvPgTCw4O/N\<;9"zXl1$-IuXvV >W$$dv]mAv ( ?JnPɡUe+ jF+CT# ŘVwCT E^eTP)>9크5[1j"l#㘌@If;vRtH5i`#Ӛya$z_46 c%6rMOM 6!nHvߐnT/U8y@)!z!Q ak _P1jMoa,{J-m=i {>_J?};cEgQ TʶͮేV1_/J>"yL_iaYjJ" z.QF4[`kىC%[t=6#6p1=NNOu]I)!)[6]"|?UQaPRǦOU:M`cBd12hO6B̂R$@  8C)cHZ$LE`QدUZVJ1\f攜f6R, 2 -e︖іPU_ H~;)3-KO܍oboّwi}Th9 ^bILU*` 2H`KBFV_2BX+\*OyLA6N^,,lB+iބbTuor) s 2b_-xdm ĺ9nQ;<^o[jsf.ѝ^|OǢtYrE:_gNe6([Bϱ&u۪هj6YY *\:.L6BL1XW ]DOp~4MGlFh#VΏ%0]ӈ@;3U3, ޣsъw?$Ӻ)T>׌+ ˖Fo,-ᔏ|Z" _[jOUЖ!E0lj*6%nL!v{ݲo7ny-g gcM9y$yH/_-n7o] *c^QgCmخ}m#{h5 )G%xk'k`IrwYxBR(#A}c\pL%Xqm[B/Z5!Z@SC SƱ pO+޼k"Tq$Jɼ<WQurrIѷ(Jj 1FT;ai^nxmU:,X힝]lvv. _+C+cߕaV8$p'Ėߖaߖa|p][T:g,C'x4m?KM>Ӗ@Z'e̷ ݷN = LfL&姙LW`kOӵ:+-tn4^>]kkMw<|!omnz`aзbr> P(Jd׵?{֑.pUU#3} A2@^v!RD]LQv&[MQŋLP[:d밻D~2ChpoB.F|Hq߷t1Σv^$eh@;9;@jΊTvE v;r kDSEusu<E}y<tv`9O=k}p!m0sDhY!y>=>~' S 9ot}BTZa+D`j<6[7Ih^Ds ٤lWN3ňvw226XϑD]<,: ˼5w11Q \ρ*{L+?Pf<ҲԦUzm0PefYܣ^w̽ 8E \8y4a[<-E`\V X g#{|_f3ꌢK XVmHQbr+,* D^\N(D,m x ɆmO?Ṉn=yȲɍy!5\`$D%yQPԓ(Z[IilxEɚmd&*٬ VƷr%z^˨wJb=Z(4$pαV7߫58JbÃ?K]'}ŝGʚ}V x,~Y"2t9O .U>hWN_^zo@?oj#5F+r@G>rӬv{l&-CY ].DE9vt^ )ֲ}V%k# dr#o\mкD>Wj^o.3GWgȦT\1E",r%F5bDqG H.$b:`Ȑqja[}-Q+! dyEɊd`Sr+b֮ۜ*9eJp.%8T{zQ_?]LX8r 1Mek Dp?wVb$X64sT[j IȢkZkz$пG/|Rx$%Ѹ??ZP xu'tE9j<FGL`"pb(!J&J'|E+%q2j4egL=[$b [H2o''Qﴌ[WBB {v4Ĺ=?wg=" 7kmԪ GAFxZˏŊurB=.h&fMl"YiJ;UCJ(Ws>$nEiH'$TY-X1)v k'2\ٲOb ˶虅]1hȯPvju$3S+ K>I3L#ې6 9-x"'z3IOiN}\5oSz HLq=эF販ҼjIPut˝nr'ofru3pe"(d2:z^wF;1xWP4iLGk%Mha\͵,[] 9G~.AZ,55_I~E5paf\ߌD8|)]g< xpV|F(frb_2.%gJ2UyCM6bԞʿ:4-/ǔ3:Qt ɠ^T zo [&#N s#@#Aiq^ JN?so`,?}kja`EJ^+I{]v}j_Ԉ>]W~e(ť}3)Q%V4Jm[hs.$m%٦,tiRI<-![J(Pe[ H)7UXpX5B#A3MfC@ k{>G\>vKXh} Ϛ tv'n($GEVk8*JtTdZ]%ˊzrj8*ysG%Fm%-tgGw /,[3Ya_/ۊtE+֠#q}kldoZ"قj%neZi#C}JB6^ZVR5ɟ>/zhhןq2.P^+i?V;D⃇;4HZY"#XmW=RPrmVfec\dRy_maSf U6d'ʦX'NQq;v,QE1afI1HFg]c4ZVF d· l^U6~{,{?KiwJFѴ[vYpc9ЮyId,ߊD<&fH8EF4e2LV7Rk~9iDo F[1jkaYSChE='T-_a^Rak3[Y\Z칟L920tœR9:.:,utb,&Vx1FA ހ`PW.,xk2ѐbÊؖm)QW'ѬE9[L]%k@Ɗд"i>r*5&^sZ=NV 7\*UW泟k3S OJMN\W}֨Gj-L{MmKt[Lt+vv}6{ AP7Em\]P\ebgm*mj=aA%ٻ9rWXHJi`a}Z :}*ӿ~ȪȌjw9lGf"ȏHTU$Bii3=bj̬Nk\"9m +s爉*b1bPl(ԔDٙ@n+sh=fXzb%F4 mҷuZ$:xVUXBK" |L!)H]ȚXDdr p#ʬv5uY=P)D>)7 #8[RlN#WPUk'[^$ү$>4ɦfl1^a;dit2Aam=fP}N"Ѽ\H^_/ejGJofI!_P f0B$K0+=&߃;DpL :";LjLꒇce)XxchƓ9:5m)B/yI*I ,T;-{M}ӍW fH %1T$MjaꘔR^d ;2K1.KC*A!*ΤZ =g?kVqV5F)#͠k^\CbMKBwc`m׿UHGv;$T ڒLL׎/އ9G"6%߭a8uN)L *Iɡ.n8emvH^_lQLJ&ߒ"d0߄ʧFs4QF\Ҫsu7a*沄?ӪJM x4N .2SP 0Nk@sDž `ˆ'DV?=lU"ow\;V & nw:ObPl6Ӌ R%lug|jDj>k%$ta`ҽ[b+&`˂-H_w*7ZǰRe BU1Bs^|Gً#ԛnPKs3ČV؄O^؉(Sj7գI6C SB])u.$%zBs\Q_:^sKY14<%[8H$ 4n*+(;!|!iEK79Jj>Hۓ@=;%FTFҍxJF(xrbM]%%Yv[vrѝ1O-֯IzLМQL!9ٶQydMݧ!?yVј+ñT#o(+~#ŊANԑ{ ́9a3Y v&xNYuO&!ւwsWa ƯҪpyl]BF-umK+E#{}a)n/ \vEUJ'[޺* ApTScUz*L&FƬ.{PaM?xFݿmr*6 Pez kaUvVqP%B.n )Ԝ"kr;R Iٵ,l b 4 pMbT딣15܅R)?ڈ"$*"@$3%Ϝ<"ҠZ (w)Y{\$p+R/ W<^f'+;#B{wZO|f[D˗*7 Z#Ҵ!t/l8qjQrU*]N+%RTHݘUo?VeI=X|fWZ?7Y>ce .v)6ʥ cۂUYjV}C"ʯs2y6*\-\/\J" IxatoUFJsXsh>h' W6Wg~n Wm5(wF}wQzM]y/8b֖F:NV"fh.}TRto{V*Ck؏guRg220痨QXAk](,rBr\PcZCJn`BtDѱ)$ꥊ5J]$Sr^߇_s ީF`P Ӭe;+PΆ|5)Pd@5:QR:M"V]@6 ZH 5нW-DY6:ytlXk{JEmͤ$fn T&eQܢ~C4S3$wej)r#ajNXtyhVEpUZ*8BB.RFV _|ǒH >MؒR;Ƨ,vdO|q$Xdw;a8t#ĔԜ˳:vJ"YԒ<J_FPm̥)j6_teQ+F'׈9(k=NoϙߞQ>T2o../vV+լ\6mdcL<_b% T9v$yQ՞ M& f,TIhŻrh{6{\QCDՠ85 jyQ\P:CH&P3B u|bJN7=n(Q[ n/)7t!n]28b;臙7tw_4j2iT 'V}a;|:mXp2̓HT۪?ObJ r]#?|ĽG<'?a' #hC_È։FOΉ'. 6r^È-;r9E?o /k9Ǡ\~W(^ų;\{ųs_xGXn7sG%qƕbϸ^_l8TM{P&77D2*/ KL+91Er'R<<j#JczCꅺG !' HSUN)|;(TU:7бlxuGlj5qA'@k#%,f>=v{*&|A, :Z&"f(CgAN?<-Z~dKh F9aRp}nkʈ[#D˩]f yl؈`$HcD4vHT{oG/5%9P}ZեV}r17W<˪_qL] іK `qaOtly v#V_hW>oṾNwNcpw ! V&!y/_DO w8d9&~d\|$Kd9xSFߎF`0:3( =R~3_m1_Ӄݚv~\BwdxM*UM `?GPE/Y9lrVL?o>8.RRPi;otNV,>qMJ_z>^;>+|Mª_R׶DKžVdvyIf^6ɻ =,`X/mNWLDhz~/}"|kov;4̨h`U; .sбmD\6pt:A.%ii%궔l>hv1XWp(G3.bp/|v6xH%R8p![.VuW(Bmp)H,x4TmUM!j{>8:BVzH W=$c4ƃu?`>şh]<$Ir6a(R|f[|Q xYoR gv=g؅[D@JAjV9a !B}RĦւH:ʔpjuHn+*5f=GJy1E @LI,{#gPpFAaXY Gy+SlmIۙfa*qy}=F!7Sr`ws*Ol<+V&lԨP\QxB}vlSuj \\$]..7n*M ER ϤNbb)҉ pA+q xMj`hsooQ_?mkPdYIs?~,օ Yu*JjIkRw1s5fisꕢsLUs"UubI?%`߼3YRZk'[[/c\U) *0EmE"UGXAk3FA.*QCѸ;V)F$+K8b@T|^fpEw46U"h&a)T*Sf;f)!9v ޖ2%w{fH>a.>eR=;\vOʡVS=7jњ3RB)<ǒ(BkCiΊ>!zDbyDEHTž)l*NqOQsT Qڡ(~ȆwBZ3ʞDJ$c&%xyN7vᐚXls;x&(8Wq3k \FJ\\]4Hejdf(FT;eBCQ*cuB`oCE ;R8E{BTsq*Zjj v_Ŗ)դwY0b*1ӳ52+⪺la(.d"窔5 wl Va"zmୢ<eDQǵ[*\Ӹ\ۺ'7IA$q:?}=J]sgۃR^LHJO^DLr\RJvߐ 'H>7CDq~uQn5FB@sMuC$eJJ*r+F惡-dm\NRkAM&$*r&6d%mSKcrvE1#]Sl^uP%IyY@N]B+J&W]=C;F[8;6怬 y9QtSdxFȉgi'"U+_L=gQ`)Ɉ&(F!$4TW T0m1r2Iq Z .ɖǨ6skHcᩨԠR39%Zϗ8̑N]1fc@ <#oRDB ʹ4I$tU|@MRUUP(ы }<+o뇫(]Rɏj΢k i5YKw(;:6~?@QqtT+Ń-/B?^Q\eBj˜xPWbH=J22q9ډsϻJUJ%bQy_!8Kw+պ+QY3zL:w5.?b3 y!6P ։8u eo閐QX=:E[}nU|q^ġh ߆kS*[mklP3@vB-9UJNdkCWUEUQ> b/zhQnj,Kh05j|4~"&C1ƝjodZjR"*brL.PZ)SUS=uy]W8*H*hlk2;=G%4.~;8s% laIoFjḽ9oQjԲi T4HGuQ9Ĺ8cgR^8SUܣjb309<'y6`2G6v`r;„'A.S'0#3z 4xAOi,ۙLgC1&roA8SϣN(ܟn9d!  &s.}Wa߮6mHGDT-Q7դܱ"HsQPK-SfJDbbkC }R{Bx\I6$>Ui%ui#8׬vTq5łzC.ϖ$)*eZPELZfTQZa-ﭕ1>ENzuO#B)#A\zAyG0j!Hb'oUݩ.S蛚ԃں%,kʨfŭt8kFQyMqGwA~t6'S(&sҍ\»8'[,QJXCkDS 6^6d<\ׯ^)S-ͱ+YŐæ[l/j]S}zUy(UOTa'3vFI˃*3>N1t)]!Wz6+p!tFQōZv:dvBNXaCl"6f"UQI%B"?J"u!WP,]mWNkrY (d>o>EvOy>z QOUyw9M 7.Z9●6Eك`zMQ֨ӎwhsbFᣢis^{[ٜ17 }) lAgh2\x9,-a,+t1d!#G,qgOt|>ڒ6 tXP9{ Ŀb >QdbQwݟgn8LXq3!uO;Qg[?x*'|zͺŧ{kUIm ֘sӝ[C^ R5<4UJQ r♢jtɄHm6q2)-/{oM _iWw*[V}s~8rI:=_nIY{痞L%ռǟ(OO~}se :UeHts-1+s)Uޯ +}OO+vE\ _Qf)? s;sS܎_lҠ4[s6\[UoOm]~tG:z~ov4PQ8?;rcع@DvUHdܫj EI;>^|VL?(MfŰZWkއ%}rCx [?1ѷDFi]<<ŘkJ9V˹.ٙw}E<Vk{vr|yG_TѼ?&zw/<_{޶e/12EA\.m˖$),IYDY(ɲ"ubGi>%P(ŷtvq +)Ti <*%hex٭xJ>B'e tڇk}wdsf^H%'ZRVzפ[q1V50MZluVZF*{p@fh1j,˴TI` C3ױaÎnl3C aǢ :%[&c8f([b27 횵_J5j7h*:X=ެX9'៉[AMvް /KJPo=*c?Xv_ B l_ۙ<`(xܻEHqTflǵ!(uכJ8!e0%#sc fiȺŊ!hS % F WGhk܈xUnz^]jk@(xyQ!ԃ ]f(9Yx3PMߴsf4]u6RDqD%h^M]n|JvI. PGQ$^(U(ΒUX/p :STZ.VմtlFw}jբ >J[HTW-Q\1EvؿO}W 0x -H9SreP;Q0G ɂ`;7л³`g:,~*Vջ˿^|zoOv7k ^JPXa=3i/M؞R`{+]Y\12 !džhOOUS[PQX#o_ߌJEyfGZ|D!f",<{m x~*Q^=`9e׋%mnF%8i:b^s(G6!>)\uf| +ۅr%~zq+κX{ț1-Œ@{  Wi^a_-kd)>aU'uxm߷ZFs]2rNFvaYr{?ɲ6(W @xҸg8OW$#dO ˅W}_VJJXD"ِai C46͆s31k !B]_FBti-I r n0Z.%8(l *c'V7y8|6}<jTOON( z?)j-{$e9H3x;{|EFe{ C 7y$g/ok8U&oa6wM5זhT iuV 6E^pLE SxECA?]TY.-dV &qj"P*t.v+BYCM('XXXH=(K1>E&P;:#@*_ʅEU%e/"cRP?wx=oh>HX1Z&vBcV26qJꤌ<̒p(&LIJ">4nh Ahlz_dU Ȫ1t"D@ Bה" zXT!@p{F'HfXr.i”tnZUR``g n[UIy H wLudnF3&YM@zqbHQNg|ZI< ģPI< ģr%˼h8ϛnp0)~ t XH$Ltȫ؊ԃ i(R&X:JeFB 2MUAQȁF`8fgQ\ hMcńn[=lÓ@YIKb['GrQ9GƁ(I#FL|<ME*%屗 QobʜH V{]Cx8'SP-MqX h,ODVYNz"OpEL9xf7; f)tB(Iy5H DѰ.jzx2K:E捗I"DNkf7 l!QEfdHA[ZIԏlԨ6$7"y jz%CVLz*+fcsRo$QQAvt2uY?]paDߛs#6ȻV\eU~dBCK((3AUP7s[Q0ty`L| ྟq2aN@*34 P /('T+L}+B?zer7Ed!άq3x*# y$=MLxR_Fn^[;ߔlU,&aS?]SR5Op;khXR2a=uM::/Z@}?Js5;\zХSzb9+@KI"c~zi#GQdb[sOT %0e̲;K~j*н#a< {Bz"bM;"LQr:~۫O8|?AT U|]JЬa ]B9d);Xc+S^ᰮ~mnT.,B۠BD4ZsD41+GDtZz*":>04t)ײRFjʪV|jF(<_h<TQ! h׊v>ET碞\wBʱ݁s&_Bg}&!NK5 nxYQet,ka1|,ʶ*>r$FJEUVunUJ&O߸/=T"<RGVF J4 \_1W(q:T{]á+IIo}d;}&Q=})hphqAVbeY(kmFh"ʯF U;d;_Ԛ[cO1S̶@}aR}ED;bEِ._!t2Bͨ2l}5G_EN+_ 旡e`~Y`^ Fށ)%Vh+J0MNXPUYf*yOMINNߧΒ笢eCe P Q$R]H%'ߚ5b]BsYM~~)nؗ ܪXā.Z f&r "Q) [fkiT}e <k|A9~4=:JIfFKs~w{O`Dg0[)0bd(ƪ=wkpb<3wwb6@@Qx"81w0N4ъ9е1ztktiӺy2Pujy4qMANrAzM\3ntE‡D2XzsceDI\f'jBxwId_w$*cȹn}TpnY{I/u }р{K"JX(t_܃xKP#ٟwnb%դuե:'r:㸹XƍR;o^AsFP"Fh"Gi k %, CH}D?/ܱ1gri^%0t+4wlķ"GEJӅ{cROcmJHY{݇&0ĞZ[f81Jz 2k!#Q/h(W7..p8BD} 0|}3Ofx6Pҙpq3f~~n?u_On"?ZdᦛhF zѧغdX1JakOh*aHFww_{۟\c.VӄLRKK//쯳ne8:b4OCI/qلH+p ŚU4 8Eޢ)l ;ִd%tHՖ]P%B󐪴6B=?lǭ@1y9;-/9ձOYMlaQeܡӺ&RO 0nPr6=yF( z:m-DYkIQ' qЕl}O'C~1f0Gb`{5 Ǘz%I@aظ؈ Z*%K%UdZa5b}#7{s,!>á\ϷA|8b0{sI+Xm%OSÖ.sr@*CGw}oC5ag0^g* 7$4&[X:CApRgbnSaj4eZ/߼7EBQ]:[6G=y;XQXQcEnƧM?9G"D9!(5ZXl[cSbF<Dy 15#DR\P6j `sp͢g]H½u\}a l2]AئWW$(xpϒR4a ý''N 8 y]Eͅ򮕲w~2q;Pv_LE^ZoBoFZN49m2Pvo(;7l0 {olm5l '|b'|b[(T9 L13Z'&##RQƼ#π_[Cu|c){mVJ{ܛ]KmO(Ẇ2)n!An]ʿjVQZp/Ԉ GjHhlqh]뒲R:ieO\XXXA]r0sZms98])(FAvߜjދH+ql(ڥc_[1)zg CPz-Zc׷g{t2dB`OֳDuA&S9)RƨN^rAIZGIPȣWi*6W|?q (@CDM_\/F?}zބ[{V d1zRC. &lwFSsM|yꃽURʹl9njN"KFT &0 wAӎF AHD8keffՙȵSWܱo-{c߬mq:}#ޭNߙu8H> '>L4TJ|H3ASH%-XZfER QxAB'6r.u?A .heׅF@|G}3>T:&$ЈYX Г5"c@rEޭƮa5Z<ШAZS+ P ;<&/CL4vb_ bMf9y0[Z)$HPk&%{>O={ׅ pb$ c Su9 ["K PRojzǟB㯑iYoP#/|>358g]FF^}f Lw% iTPD nS@]Pȇ)-VYgkR`OEa ؞6!H@KfPEM 5B, jH8Y7L)^WͿ9ļydx¨bjb@,9#H}De+%Kr2#Wz?cn/@ Qېc'-;ZJF;aru*x{WۛrT±7U~Yω&eihH{w6|wu GYj(uN(%xH=3'F5ZuCVP$݂w )ea9 %N;wOjۼVAj[{Yû*Q9cRLZto Йe⻙#K#9g Gv2zs}Coǟra/L,%{Dr=AlN^[.9GHJ\rVX%hwQ$&bsH 61u[o^xKi7ӊw|2k0m^h$f9 Sb⤚2c/c%Tt;{[~i>8B6sG\4+ Epx0_-N'\L݁xM =ls7^W+5W&竇U ^ll_<~U(rJTߒ]bwʆ- pz|/K@}k4@TW=n$C ׈CQܦFtICVl;{۸60PbYkB51M[TjZ.(2eֲ Fm}`s|6qeJpt&LٜB1UFW ȇ^ȃ.UJs$M)4D6W W)KujaXs1{mkeX[&/Q0p|eV~uy`lهtʳ xI @ڃ`/N/}=>UG3lM~}nlûU{40>ҧk`Ӿ/\.yY{b ;ǾnKk4vourc |x£7^ٷ|wcF=X3(|VTf].W!9yHfQ8--Aޤ\q|'Sp mv"8hި6{iú|7n/@ig̀'' ?`CQUsGnscv:TTvЦ쿌|^_AF?2?VZ߿/e^iiMA܋&; --SiUt.mHJ6-MF'| iHq14 n9^` Yܚ:~ Ut)@T{˱~L`ՁZ\7mcO#v\wSF@'uÚXUaEHV aXXЁI J iM'hRyvM&Gg'drdovpo'u߆=ɔó@I"%{j,`v=Xb+Y_Ci1…oⱠ)~{$> mPfоU|ӳvuߏcSu}1 8#(?ިCƄC@:{ƋNyӡ5?ҕ='XP?A}R8& ]1o_/WA-FL·~۲BF7B+s㉢ *~0gDs"dV0ZEX`0b1bZᙙ! ap$HK$ jMqy h,6e*?m590Զ1i GPpn<("PIxf~+Y9VqM3҃gv3_3;xflgЇхk<3Xmꘈ&bHC"g$$%J3J-TIqVy넧o /LfF5B Yx¤dͧ[vx䦌dC˳'ۿ.u :ލ#w\ٛw"̸ۉ`ن+7u 7LxUZT"P|)D9[ }$KiU&kT+q\hjzPR9QJ,^YYUʗ2^\ּcRW/2{aߨrjW6K!6ޕ釱}f-t{8M:ajȗ L *SS5v"D<wWw) }+&|w6w6[^}s⫛7v7Fs wVBK+;hmʌ y?kn_UϪhϦԢew>F8*oZ<ۡ>|y1*l :k1[}!\=rfD twn)h~x D[7/7\M;soaW*fN[Ȭd Be9ʱJ r(q8u8/~~5<::*WCt/'(VKI1HYM4P:.6^9D&<.KlQPH UC!-y%7%}%,UH`v.u%ްbp%KbDe~9+=J0 0RBkG %BA+VPg 8uU'z]IM4&_L)#_Vd}q.|@ԅzu׿!BGejX-C!M/~y˓wpy֤jb"(SZ^cUNo`nLc4MX:lAHBr{SCn';0# `wx.z{ߺ?ٚWxo tG@5Wŷfz>i+in uСq5|R+[kdʔeE\%P6M#`w߈LjHu0[mo޻21LR~zAx8oP#][s7+,f*=l9N&df m%!x{$&ٔ6ْ8-8W|g+UeKu(12ZV^8iؓ!P0WqQad2 .iaAH#ё5NqIqN~/BVB"{F8 +~v]"@15~R=D!d9Xx*6$E+ĚJZ8*1Z1g+]{9"\R+%`Ei*W1TI{=VP w!,"@@Zsd{ b[2cX%5sciV,heHE[9Q{VNgVeCY*PјKD@!E8ȌXE@B =Vta'GJ^\|7;\$xXW`HN e lT^+R*qx,2%HRQry?gV+*RZ'+M'.2`\?>lqR`r,H,* HN;8ԓht s+*\Wn(Uֱ_p^jSʂy={ #'1ƫ)#)CePͅu@O\ /Ui' weͣ艻[fT?~Q!XxG}pvz14/x`)-&jLj򷿾=k]e Yh8UԿw_\7@-;~ Ƈ -$2Nj? 31EJ,i~=[ n3Iݟ6HC.*`- .e @pM f0 Q%nk#1. rl2@ɠUR )Қr/T. GikB"r8XBG})VIVכ.a2l\MKV5%pedM5% )X [!O3 iEv}Q'[-퐑.J ˾ETo9ٿo!cU;IL`Ĵ0hQTE,Z(1;ԢI )QGEYvD`^wWC3C"4D_`TNҪД^zEJsڍǔW+j'38#< z:Կfguyl4yN966|<~rLC7|*du9︌DaD#ڵݪAHmwoNa #>'$d V {*ʊʖ 0-bW#,+ĝI%Q̥[jqi>ءĈ=M~K8a#ojn25JԱZc&aqHyjb*V%sS` 0zH }ܑPv݃WXoYHn3hDȞb<$H$PSh;rR0,f#f, +XeܶQ唅4_#x6|ly& aλdU=}*iGCQwA&=[TLJXvvUI Qү*oke76h3Ɣ^ϯ?NFIq3C 8͞|H)֬y"']IQB ]LUY`+?Ua-yLr\li 5TB PBXda=ke[c,xR=+ayEKfgJ%Cp@)H<7UȲ:+0JrD-,idK/<=#,Բaq`![g'!^ xPu"q6Λ8("&à翾 (UE*#DhLYг'?zQG/'ZJ$ & sW"iZ!ы@Db *&)F:+xDfqnhRVUa!g")A W"X`Jf++}B[Mi…;. G3oJXw!Wj^Pm)(P48\UJR(Ȯ+Mɩa 8 Ru!H4_0)q3LWi8qVr``d%xA٭d"z.jp5m9x,-!$r'Q!]`@yBaM 2;a01;q$q! J֚"*6 )lO 1'@'kZ`Wψh$V|9 jIO9Wt#ڭ&=En'je (vq@S%)}QBa=;\ 5o[{)e{ D)A#R4OyEɊm3h"gz>Tݟ bJ7@T~ |84ZaLރu *#P7laU1{s~fI5`FrRn@Tb|_v1QX趎{MK P r/'4MD D&J*Oo~))Jȵ_7A61_o ߪ=Me-;0:l{{̎5 FS)xbco%˟*rk(_#v ˇm(=JVyYq1F]qGF(n\Pc $COa I ._! D އ$0)~ C YZ MZZ~<=NH YcdWB# HLCwLp8y\0,&p/Cʧ@DiY\JL/y'%3"z{YyMˬeDo?H$i$\j;w{& {} 38okTs2ydGm]JVc!^VH1U@a$i&I%5sk0$t%6|OWnKgfv0Lq!A"]Ϛ5n]zOD?ֽ&|q;<33 fnl~cf`#O.&~SC9w;vK+%ܙ?پ_/N\;ބ_? \O}w6%BwW 7i`uK؄ιR{!eLUNqvw">kuvXg) L#UX [+sΜAze7*kjzRLpPPo|`}3bl+gІrJ.e y6pW.? 7- I\\M'[0ӭE,cR̲eHzqKaL~X3kɽ5Bn~7on/2>,ޜc4%XNє7a;?B <BK1иm8>d?O>̠0O@mm^ouu!m-⫳e]̪/KG X2}͖Wi8A S]nH,-!H!)!1R~BpVONBI#7e gb^.bLwlw,ZZԨ \vIg.t7v2 cHu24>.pjk?X'dj緝({=SP|60s1H6jZcYiYb\^&7˪/ c2;zCϺ?Z!bo| IGZaA,\NTJ;nP21m@ ;cE!ko¥l~sHV'*y'8;ϯիO^W>@n1cНxFiS=x3ezG.lVDZ>udžVϯ/o.3#N^:!/6iaN0,<rTjop;E]#d!,wdˏ}2Ʈ(QBW I)>iEhpaVڔ%؛p9_q@Stk1wk19vx(nz ({x51;;(IfX D8[М_$xX3w|$xAskB#=xT2e\/J 9Y ~e^d*R~ #4B9 {9\@)TU36i"&HOb;QfiՓNm҄%S&גùSگ ͝rԆ`8dDk)!}Z`T( oҸG,=ܣT\6ƽ F(-ϵrMb4RX<B2x0jw᷀;MjPP+ Z^׵_3 f|iƫլVy)qt5!DVX3BeD🦌XL7w'CqamEo&% cp:`U} XW" `]$dJ)h!/ |ҘX˝FQ-Y! 'JԠgښ۸_aeNFBTgkkTRZڕ%rlocHa4C I9ry8 4]s*n9BRgkg;V-Loˇa9C6zTkrjnNtkQZ>Z8?gZP-Ҋoߔj@zwӇ惹2fF M#,&+iS}`Lds,K K9 t IF4a`L::_(`Z7\T6̳Gȓ/Nm4!QQ /lNkddFFB5׷y. SZ,V)4kW~I,GظqJ(P##Xj qk0@>ȖNޕ!s8Erü:ϓ]wCgZ*4at'XAOqbhV.>:EP=EPdeޠBZ)4YKUzƒ"myfBƈ gnK3R_m FQKbu7\<D@[ڎݖF@׊\6fL"!͙)Uڈ >TaW< LH> Yԟf̲NB,04Igd _wn)E8o_fSWj[J2η{a󒂋'q~jLko{zɑ7&x~ouX#|&d_5${džf9dl7V$?>#> 77t|sNFR7`5C@Mn Zv`C~XU-*EQPQۯ ׵{ Jj[}`zoߠe+oY&v}M`Mig@pc %4؞C R#^'GOWGW6 oa.;no0u._Mw8#->CbR᥂sk-h:akԆ {(T% LYuwnom'KƈgI΁( 2Pc#*гqQli00i`X\{X/9^l+Izfw054+ Z`俐(a_zrQ{{RK74ǫ^`p3 W;c5ۥ@nJ~kbߊ7uo.zgݛcU?J 7!?dxOGKL,$P>ZQx޹TH{"X> 3X4Zx9 ڊ_ݻx~WU+)},%7yxd^,{\p_%PW&N@0V*Gbz}K:[ҟKvKJ#Fx=0`ރ.TmVrׇ#߯ :*?JۧRBV2.|[xKG:lx#M{ c߫ {M _ S9KkK9FkL nsArrhE=tO`ΞŹUFFgϲ(` ZQXj`7o:c޹Qx!@iOʙOiA Jm1 !;6Khީe| J?~d}H(:T({<{|"u¿ ;`8ٿ-2v/ 42y*P_h n}3¥gQOOS3\ wJk *U}$;ɋpt⯘Qƞ.+%8T(W';r=~C1NRV%Il#]#:Ǜ4dRoz)syHlGq6yς\jޑepƭF%AH՜gbZW{uv/8֢G$8fPM<6kLəL 6+N#M3)%P!*#u>g/0sl&.**Ɍxf ̨ 3*wuWgo69Y РpKĵiɥ LdO$~y&6RGfQf ۣ/[CWœPEX?,+AU/P\c2HOA$T^2bk6HPHxd,d޻$y q٤haޡ)$8 xk4\>ؽDh?z$$B&Y4Ʋ)GsMsDB ̈́Pm JyOҟ(ħu`Ȑ <05C8FU*KA?V%I30&e`R3B*pq013)&Y1՟la?~$̎3,՞g\b2_20ȡ=ץىb3mټWY#T%od[`t述ۮw* +B٣YnS Cϐ׿H~|G_LBY#:g 4=h`u3 nRqz:sZZGm; *f0];^Nv*rqfgѬѯ>NjK{*;Z"J# n }ݞ7`VIɁ'ZQ)E(ihɴuj]?̦ŭ]0n./." sM;#L!8yJx3ʐxYv;3%Eb^R<}7Rsk n! <ۇ"O0oh.]A caOpo ܚ,bʰ95O;_:NgG&c|@=>4+^Y^brK{֜#D~d]fO4}>]L`6yG}?k;5C%36e_Ϯ_n=ЦWEء c'U92bUUlVIY+ -d6lһ6飏b#Ն2Ɂ<5ډ렑 bIn>wtc1% #+K83~ƣzs(~@xO9Sh}MNi`)N hRɾJ9Z$os_/w|+\뗯8' V}o[F A>8{[>bnQP-@;lk|~{g  äL:NXϗKٳPQ࡮ār*/&"NZ:&BcR#2#:2, 0:J@궾纐_/k;Vtfps)df7RVAө!vBYhVh_vB>s)a^ǹn W 䝐j1c:1NE(PݪncXgnU6d 2tyfdrY}aF"uH(m0GK1+Mg2X#u9UHm1j?$,P(xIMd̥Ot3F`g<$-!Š~t4XT9a=1,37*Rկ$&W=!btjc;1+E nՊe1,37*qn`+ 'dZ Nw aFdnՊe1,37ѽmJz5[&$) XveЮ X&7ɕyx?_.9:'%f`7(z 呯܌|?Sv)/{*/ϵQ|%a쾔-sys};KIvޅ+"5ё[qۼ֑ņW-{'|{!fO  E?Nqv8{sD?]f9-]{7ʋf2 'Z.MakrC(ZZlr75 3 m}ߪŴ[ָ7gOtdvBgC4M,oLȂXuy!f6_:|ݾANC?7bs3,r f'DK؝Mt'f8@#%wt[ֽK.5-%ѵg躺:;HUz"*T hAa1#(g40e:S|c U4JZ톦Rx7$FGubƋ#ͨwK!\EtJGM8`2`#"bhFG%Аg[:=u `1I=;tTބ0O,aM奙I vTpKK(u9 bn5K^ݓ2,ZV+SHB n A 3 h${ UnS;:CX$K5ERQpԊͪl CQ)ti>]}`Ҷ&כ‣+0*qlLATeYhL[.M@J&kf%*deU cPZ [%9wj$BT2M1fylJNò/FQ>σh9pPEV(wX$&iʭ y N kJ+HPih]'|(&\v@E{2LTZ=IcEW {Z 1 LJ;kPW$&Mb" 7Zž2.g;T :!2)l}FH>dCG)4䙫hNzyŽ?f Mh֟ 0]˳`S՛m9A&+/Uq4_Pc2;U1辅F3G$B h2V¢wY Jгݏ0!#6FS`'V@479@LxƏ0 &֍c@D:n>UyD*4kz{84K׭bVP,٨׭ :0< 8Ϡt&۞l/^g`&~=zS-7ꮡ?Ó7MYf*4+몽Q;~xS֪==%YHFr>-K 7ڂĐixX\j;1\q!WθqOE%63GX[w슠ZAdz_*}iΝbY8Su2.#92xȚGWxw}m6IIg~ M:Cפə:*yy;`#өB:Kf~a9H!mEɄ-Ҩe N[M}4Qש$ӆQ+VZ !trUPVJ]yeσlɁ6tKZ6=^gs2v2nCexHxHګHzYf32 { M'Rzg|!q] lczNa8]m 'jicj+(b;_<\zz=b"P `uiy9ʘqBbBYa0er8VSĬGPˑajc혱ckZ9yonB{#"LJ%8 e@'eI 'X)ߣ"Xj;z9Ha#r:r˴.- "ΉƥZH+a _9ߙ$"c&!>r[7L*7 GrPmWDP&4*,T 5B9QPpb r%ʵR a !N–1UPiJlʪqWQfUhk6QruzJІ\]R9I}STY8}>hA F *I"nowͪg4+ǺZXwm0a۲e5\_}TݮBdnhy{w׫\Iݪy:;uo~2ԫ.(bpft/x9ӛ#{1_,6B_OajR-\[l-o5h!"D",÷^ |APK=ĸ瞘H)V*T 9DKKB֖ Bbt Dkj u7ʡčD3\I De! i)ψIC;mq'b5\8{srE ӖwmSX VlU|LPts1Wǔ$ +?ڋpO0Fgŀ*rsz,Ü_4.f{::wn\$XuSx4ܚiXի܊w 1 q!HPtVJmJ7:]yݯiёW*MXOkU%MBDk̐l&&Ɂ_M;p}5//k"KASjfMh逢͑3EB ^NPI_6$|QߠPT'nmBw(W2xVo'G~sO]ޞ_\~guy뮌5~" ݷyP?z( Vp0[0  I'xΎ~2^*z'޵/gېuPU'  Yszj/qNՠr`2ZYvA Ax! _}H8=({y놧ƈ0_* D u .I 3{-;uuq-WZ\ra&%cԦ$0K9Bbb^g{?;fD; ]c7A{|Hsq?B{S`՜pކ)I}<浗u-Ybv׳ʅV-_K ~9nߧ+- ~>cEywͫY4?r>|]7)X^luxLpy;G%7@nw3rIJ{VTY\&Ґg1:՞8ݢT3<>O>#'P'͞>9I`B 8j՛A sҤXA=/9E9޽b N:ץGﴐ mQ _q]#\B^ ɪXi]sFWXp{y?XK9N.M⊓͖j03yK I9oHKi:"t9AZ(ֆ|3fXNzƒnN4AaЛT8鍧鿃3\^s«"3{wuLov9mr..EU/}]9X [\:)oã6l=]<~H˺2ݼ_&]UE&2-[+1ݾ\"XDzrY{"^1D5JGm%G4j-7@?K$ׅUFA1ˬGq`J2zJk'\""6۸n>68r`a$s11{FDXbbcF%xPK@> M\ Lzeb[΂eko҃oI)0Gնۢv=GYŢ_ث$8X+0Ks mZX8l @. '; 8#<3/zϻЮͮ E(SP[xo{MA[9WrG&hRG &FؘJl5ȂJLp\kvR,\]-B?{q0|q2u-5שL4(.1.'&y$X% |V#\}'r'y ,6<Mo7=Gxw\"TT'QT>|&8ÒwԊl=ž ĒHT >E; Q&;8E~&h>jz6*DTdLpD EBFUjJcJZ ' \$VBn+EJ䀹;Kp!$·~ F/_˘P` ؀CDZ‰<1"1K 8A,J\8mSjg1-ϐB18&i*Aj:p1ĉ'cı8\xCPoBNN0;ʱR병h0y8Q;.r`pMm N%r00F}&>h*U ^Tb\eOPqSQ*4eƶOX5.#5jq9ʴ8/L S!!ZB;'iH 9(KMmPTje6ݾ M3p {zVl|>ooCLq3pg_0FMۃ6»p@aon'=n0"7`ǥaI8q]BaX7-FѪRMب7F6FcD'> gWRV 6a!vvi~$cfֲH4Z?B{ F@5x/ެCᨚl>\uu 13#MGmZ.m]7*w+ 7hπҨ&3EL)̚C9Wf"Ut-&6\:@!l#mfpS`w,WM^{G>=h y3{k}P,)9M,Hfe"f%J3nKrL^fֱ [ƾ^/K*Yx1~˽ jlsm0j&nàT"43^Oʯw2b~1EENjǴ)Ogc0DpzH/bX[p2+rBlju~q!v2F΋e=mvf ]J*͢Qj`貣lob?G޲Awfz  8Ȟݥc0 4tDˉoQYXiF /`/%(C4#Tp a0>(+?J+)n|ZfjXe{[N&A#DEmW}sd9z_F]qv aJq|c!*PͼCA`+cdD`~6@btOſeAIP;tv;+n`}ؼvUfAX 48"ҵe Y=n=zi<)Iydn^nW,>fP8<$pm~B5QT3uIUʍc_%sS$ Y_јYنሒ! m\NtWsScl1Z Kp*Dή}MCoB7Nަbya//V#2 v9bmrӉQ)1J+ub'f¦{ @Z,u'FKqø۳ӧH&V7wO^.jh<_p_f-`ɗdH-%EHPhE)-3}Кn[On[zxY+?(LkW (Fa`qуvrуze8bZ4FcS](&-]w)ֈ6ҕwϚ+%Ouҳg Fy]6:yւPO$k eZ!c$3L8Dkð%f y3ė۳"srҲQ2? 0c8%}ilnr3 -VB"oylVoZa&~xYs0w&Le҃k" fM>< _]/ZKq!=Fr4T#!)%;kX7 u+˃sݔ)74. 2*DRM2.A08(Yĵ]T{ ̓R77S Ϙ[ϴi8\XʘEgS4eND"1ibO$2v=Pn>}{"zDROb. 8$C5;cMCTpKxLO$2W _aFIRf*/Ku.()pU)ָ"R@^_/?~eMFs0 !sc3o!pͣ}ѡVr<-l!; ah%O`.&x1 ?Iq,i@$j҇Q:y?d<؁U,2q MwgNYHst&IZX]FPޤ;P]JܬH<-݋f6&QlxV^R6Qy.nW1 ` hNxj#WRͪs`.Ǥ8}LvF}t,Sjg5+P]@ ݸ?>*;V!s7!w\?0FY. o_ \8 ͸}v1jdՊb}1}W>蹛OsUgB*}מΙtf: 3Cw9E߽ Ϊ "co8r- *s)ؽ3z*c* "z. 5H/_ t#Z̥qh ֤Ю7L`4fu7ow[&8@N}VD? Zf Ӝ|nO=AuHjiEյ~{!3 ;W_y|_^8QeHr$˟oeCJ"V#ۨ騕[LUʭU_:`]uH^_7VP';&ۦ7/JP*޵$E%!\n䰃5&eoѴSBv`&)z"g &1%Kvc)$uUp8t4%4vX[D;dRnYZ93P76bؓC6dA̍}TokS5fm)-13` <ͣPzec TpkC0Kђl4 CM~S#8DB/x,osgR3ՇmJT'AmɄg܂Z߹/S/)t)!M[-X cL' S~zOO/ZX{3>duuͱV<*c͋`wfz~4 o}knni2cWC&҂yƣ R!80c ?~~?$So_C?9.D&5|+Xu^mɼ!<[ƧI'C<){GQq?bR{:~?,O|t?65&7z+/]߃?,w~Ar=)~^EH0 yE(Jn=XGT)9&,gaHi$xX}e܈|/.byj2 [S/s.s*}2C <[ˤ_@-Ɩѿl3|LCDr vEbĩD6$y^'VSuKҩ_ˏ&l |\ư"̟dI,eMkV Y5 wiĈr-*w4IMnS}zL84O ~ `&6k="\3/+Ɤ057௤OVJ%c"?(R@7[V=AKf &vvkBPjx}1{5Vu8!<`Bp}_wUIYT~TG;+_6ZM0ʺ pp,8UST+N1T~1NI8u*|,8%!Na˩h/.nTG5DZw&[Ooa,[8' lsūЊIv9xq5Y_ ?^ 4Ԣz<{&MYr qjM %go,GbW3 snMQqn͉kZ৴2-AWi>A` R+P 'B,mme L {x>jHxoGNhM'6)'e+?Ty;^Q}M#J$O[.+/mįw\JI0sbPbg l}"oI-s(0)(XL}]2F!ic"B(g'yڏDp / vj k^^ N+)6'9ug^EDHJc4_,B/(:9Rی|_lbd/|lFbyxN΍|:Oo=M$Fb"#FK$MGXM@T{RfRzl7R銷:yl$@Kpo|C!(@̞s&Bx<^бM!4¯=X\s!B? R |Dʄ!LB9>Qh`(-G^ytb$VTR dCIC#L#ad>E^(!"a>٧eD"TZ  !<0!C\ŷdM4۱O*'1MGZ024C qS%  F"`vpc|Kך~elH5ڳJG~O[^EdQhkgV[,nI %bRAҕњ{ބ [ivo AH ER>a Ǟdd}f!(wlP:"PNXimgƋO.D#X%~1=U佊z ڊz`;Sǂke6/L"da"F1"\^|I4]51[uNrAՉw2ԅZ,;jÉO-dxQB"]Jg/ Csb&bƗG/j@faa%0M&$*b{E %'/<ԈH4 Q_P/W ccǶ^: CLY9|QNEj7ףqGGF6E9}j$ȊDwMI7!繙Hdi ql{4&;2SDbٽt,M0D&x b2Ws0Qq39AT8GMVXxॊU:󉒜 YaGU8lzL0.E8 0+0ZoAW#]*0Z6W|X]!ތ:[n**5;/AElH|їͭg\'|>`{{\ 3%ZpZh͉tq֛O+Օ(&|ro-aJҷ7wѥlzgfQ;A#$Hu@wDJ[O6H|*0B^Ψ1Ign \ 2k.!T3z-3cZހ F80vgR$V:t"{\!\r'7'nBZe+>i>{%ml5.XC/x{##$͝k%Yc@ Q`pO DH41{($0'k8'`}*?ٟ9?f|KtQĔrԱݨ:i4wg]jWIcQ2nmM'XT_Nᘤbɥ#羋8H q3/5wbC gXjG;{]c/^ 5c}wctuݻRh哜㥡7YPDq{͑oR*A+T)¥tP#']9c\moJ\PZo(u^IISR׎r$|ގԬo't/q[^rL ݕrHmIaIsO #wvs?yy@i߁ҐuсlY8/~F#FΪOؕi}qMu LTV}_[`f\\3ZWD VpK߭G YjàRZ-[Dl+y?(3 _lMzNdq;7)X|ͤ);ی\qh-9(:!qPM^.K}T˞k rା1BG Z;٨OEQWگ;ߑR:͎p3K )$0LG~Ͳlj\zdywќzٽ&fJW}T]'XhM;-5+uЩeMw7qfwnwkvT?y+d_<S`XNTmfΖ@rݴl^THB:u1^P1: [db[^0UqM^) %kE$̹,y2'Cx5gpzNmY& l3S  yx%KV5!Sl9|;'nVJfK_[P8V.ԩ/ݰ|f-QstOxAϪaXE~;2Θ$/)`Xj `ٚ#J{YwF'|V\&Kƈņpa+(}T5MK@4*_&ksD먔pw2QVDxX@߂r؜53~לZIZ+)c1AZ0uijBAbu 6`,Et`$:Tu&[R(4y=Tؠ&ȥTxQxȈ N8R;1޶W^5g_3C@ͧjE6Wd H"ܾsLAk"/L0&)qho܂$snn7%nGA jN'~y|{{hAv1AU8﷑bB1d@g6E븇_ =+G#`w~Nm JyUVܩ YV&BSZs4=Z`s]Yx;QN4;c(Ӵ| 9i D;dB9nVQZNpU4p>~ݸ0b\ ߹-s'q}F^f"Gȶ4$iKX_#:d;Ug0\$)tҁQ]\|O9vL=D]CutpyB_'aAyU*#D tV#$*axf̽]#\+{1w-, "lH 1ʋ!/Gm,%fDfȍ c،wo,@V^y4jcku|n F`GqqL=)慑Eeٌa<1Qȷ1pOG{SFLИ#+}w2/ 7@tcb'phI,vЂC _Rpl=? mv?YƶD8y/|{ 9V۷T6z T륲TbĒ}o]/{&I~ݘ>WTp4ͦ$L̢e<|嘖OR |ǦωwP+$v4szwbtwў͎]X>g`WTY3HȽ:-xK.U9>/)zd(0;*ILJ9kZNkum 5Ve!q _vxBq ZR\\cT-n$ߴΌj@,zUXDQj!,twt1]Q+/V_|X//9 ;Բ3i:Nr@k.,iEc^aDr1^%g5Q:%/8|̩]R(\pbJ@N$>k /dq*k}Ľ0Mc3_ c>!m'P*. AD1&wSBåߟOe2~MYk;,ﯮ9n3oʅlLuEx!^c bn|鍧$V0 瓱Nl(dOm'wz_=,9&=l  P l|Mĸ4WiyM@Ōl'PH"7D!5 pfEBqRߤ^󩯉>#mvoݫcC[5&U"щUDE޽n{xI ڏ5䷙JW#OYo:ݧ?[]%GZ?`;;cX Mmo=-J#Jr3vn[G*)";A?ߍKo! {6y&&g-d5@Gs({Cz3(ؘwO7J>S_mvukƎO3:q3K?BBeE|9),V~1R(@jDP-_A($3r*$3< !&`W((}v$D*)Ffڴ|TGCaHo6΋fdԀpkZHX7U+[0|v"FzǏ_ 3&4/ͪëaAՙlGGV6^VT5W;2zOϪ/)>E,x5?tdjOd 1=ޯ$un⪣$6C829M3w7<2 Ã40ʯu@Z@'k1'%0'8n:R g=)N|>1,4xOm_`9 JrWhQbJ#?Y3wW> ct3OF 9!nQt{.U6e>t,W+$ nWja7Rkݍ27K=@]Ԛ%[|62[WF‘ݒrK*k &&ńӮCO( mu;Y^ &坩tW=PkYLc`Ud0Z,gF0ae[u7N23ƩaRSyLãtu9FR:2mOn919Ԏ㜷uJBQ;訽?(h>-Z:|$eKT lLiG)<񪚴B!@!gql:x x;Nm*x'ᥜzɧ# o_P%ʅcR!{ >YݝҚ UÉAQјE_ՕD+NHs:ھ:Q)poc[://&XFG61:(Z#>W 81xr׀ !wg)6zw~b˺88J{"?û&1|!*Wﭳee"'] ⏦wr"0qut3`K$y.Loz3~[, wYAzj۲"hmuPoS.TV.T4W|F%[Vj/͎k@  -M.~2.ٸ)78l!zT}-Wm\~w~b⎨ Ezn,IY6Z1)Ï>90팘HK1vF\0ZwA3%9YiI= ~32)nXGZ;{ۙ 36l*3:ݠdr 7]2ɮ}/I?Z yÞo$nK{o}V1KnXOΟn '3^$l;2و'q#?xύ7,!8nLSl ygΆSks BgݡtћW݊hKZs!t&a}¶®T&J=޼,__c*!p/umO?m놶Qh5ϞV~GkAm{rӯ;f`>h#3_糰N6ucO.B=Fl]K+QN ?4anb^z;2-BW,K ̋/a1ȶ)v̍}0Am2rnMgd#Wzd,s̚$BǗkJk;Ekɼ"9"s C[Wj<*+f{p:Ђ\ESs+  |=_c43Z<8ЧR. 6\QiA$!Mi9ϙ\ ((zx2RZ >xU"q$1aj@NrZ6ScχZ"ҞuTT*"*)Fq81\P3fpP,Y(4 1ǻ&no8ŧGұlhLٿP8`)3"!c5ְkN1#̩@\9pcMI@+m"D`0`6Fg `) EK*Cvja,b VuHEQIOogm&踃vDFB:k_S\o|;s(ma*&q`Ba6RiH PFB19z*U*=ft!*"*t@@e|!xj+%`nt Sͨ|tE|@s091Xco77eS5I\Sp]S!=hk7W޴8)U{5anfÄfv:v'z\z>?A9+Ǝ)r[Ĕje-W7\c+J5\*VQ{U#[8}t‚<^mwc(YZTw)‘Ltio*@})b$3aLP( c:ji5-Ińp96aa3RF)T:C](ͥz՘b$$<]X%Ina3v޳I.#_ . ʋTͧ0xcKZzՊs8UIN$ +X.e-T}ͳu#sO+H%9;jq}XEMjB+iJ. hTHA7cۼ}*7ٌҎWXF a7]-g5Krv5b"+%fu6=JQozw&lrs*4ip݀Ȓ6y#7)z2S+  R+ Ikv0ECxHcFO{$ ^J[v0P5j̪#3.X ^5xը`Jyr_:b%1\y<]Zr|p.piLI:&)I:vql;jqߋ"=Zӑ<АHE,m{'+ZI }e4xGBȀ  3$FA /B(! 7)Ovi.սRM CTIv$R4ByT?ƒN?{WǑJ/]|bȰlÎA)kMFV7"ꬋaPWeEGfdc4ǩS&SM aB2X#!)g}VqL!R@MF*.A4Sv&Y5s%imgwɫVoO4hPq ig$tT6wh|.4!iM# eG sfS,t-ef(FN1Tᝇ(Bx!~)D-:TT׫ܚq"C}nDfXC6W>M WRjʈe.9x$K1SaB~l nP Laݝ;*ZXm_2'ӝvY{rc+H(A5g)4.0k=gGipc->=ұ*seUb:#s9Ȝx[& 6;׻59­|ykpxRle_I.B x j.פlSs~F?|a$Jjz*{Ŭ't^_?I:P1cqi+YXAZ~_MaHikO}MQ,:jr˺)ފn~n)S3GvoengV4,!uԂ^Sz|5\~Ԩyn?ҢbahB;ѢR1*hD8W-zwG@KԲ&R^4Ӆ1"HO:HM]#]jT,"5=&I{N,J9pLGyD$l0+w脱 ." 2עb {.U1m9%Zanȳsڬ44f)3"t#Pي)NЈ܏gC#;$IJX-pzV/orymEʜjy?8 N/V.xp;7^r g=^*t;(m޼2Ik=Xˁ"EVHڼ]9\D~b"2 ߘa?;aN?UV!1j93Fsu1g=^*]v˓_n9nƤE5;nj4k DtJAwktO9_7I>AӍFc-٤2-S[Azj۵Gc'6:7njuZrp%W;8J[UiuVE (vL>ZyM&Z:p訁aKDՐg#~zp+Q% ˦وn'cDIkOvЈ/mͫـpRk9u ,tkkLbj yvxţl;~zpH1X(&&[js{ N2PpӸv4J,~+:0;9oז/Y1kA%-;bU & <߰1hmX6Z G΅3hW3xUѝkˠQ} 9Yԫ@pFT*:2*:ݷ_1LZ^F*^n7F>ӫxDMrV%1yZ6\YO:Fhng|?:zzp1k5/kd֛!t捺՛pfY={j$PZ5bI˹H2BTҧߜ9ΤJ s̨M,bR"^Z=ht1RIUI0%@m$eUy ”TvQn(E< # M*"BdX%^9' R !ۂ 'DQsVByበB c?c<&LP- ǒF-C'0rT L1nhdDEq-9B( rJ ir^K-4o H. 8;]+9^]Hǫq_Wb~b_:a*NףD)ߓ^~P{Ni"O/n9"7㻃Y\,/Wϫ;wNMw7 OnR)>7oNaj if>V[LE +P'As/Z-Fq ц+蝫x?NOnVWB:WM͔JwҕWkF{xb,%cʀtTs! T |)0.T; U,S}h Ŵ~FSB/! TXaR`m,9*dDȢBqdHpB)T5RhK I+ŗO5\]WzTKIԾ;E"ridTDc*x 95&(Rk鎒Yk ޞRh E Drg0BQ@a'KE9X܂rݧfRM 5ߙ_{h]\3i2n I*-ӟ@&u+j*)^ժJ/j;7i"udcXiUB2a.`iw嚣e\-WIH)(ò0@5?22\j2[rrm+ph? !QHLƳt ;cN Є)En3}D1tu8K ӣg*BJ6h6\xtuD l 2TlOhs3!&u$R{EꎥIEv I%m9۟'mgS~Kgc/Tº>%?ÏUY\m:>N-tTWV*rȥiA%eEvwbv\wR̓t40Go~37' f7IYl?V@ܬf>+&x`B;f>/x`HL:f>ϡ=|]0O)ZbǘgomsX.`gy~enKk!S>5g8gy:=L(ts&WQ +!=dASvYot)r߆( o<}XZe v?E2Q2f4 W;BM~ޓu샂d]9g4M-v友/"H_ˋLL^_F#[^O,(xFu%J%=-19{'a; Y>9CḼz.AY'izUjZ3N-6%Xv^cME9Q7 \J>,' c7DX/0]O{m[Q{&DB픤r*i/x09C0Q)NL5$Bi?Kת*1wt%WPӤof>yZEJjH߹Z2$40^<:NGi8-6|WFSOl4""uX%(LJUH|x|;{\vřS?+Ǫ,M,j uB5ZUߜ>GX}-|Vf9jE31%6_dg=Nvng,-0E>ԯI`\̉f7 SAG`LYidP Ԇ`oc!eFpCdHELhY޼~\jRvMbi#Il,3a<\b`:3EFZxnL& SFi0U<8\ )gJDYEDa<&#i@ i!P`(R0N+nhZkNu4HZ[\/ӭcH\~[uRjRH_Yjf˫*߼xx_ Gw_|; Jjσݳtu :K4JC<@:[mN\j(yt Κ砀kͦ=i ̭sSz] P["WڢE_,ErD*DHފs^. kKF+yqRV#HRE&1"ۖ{zme!(A(z?ⶍ z餀$6Ѹz=eWa=wh 7',Fo揢d6~D/ z U&m%O[kD xpa[oڐ[~9PZ5{maOAы [!ZL;t9@ +Yy0.5 XFip ri8E[IՑk?!MJAGDt =,o#E)j-. /-k.Fm%ɟEixe-d0Fx`JT MK&:p5kW0ј1":JK]"@MYb5,ʐg Cb4瘂2iU!f ,r2z:BܧA3OȐsTk 9qZ?sl)d>ꚑu'3tR%p=9К t2HjwP4aG1Gt^?` f?r٧mT$۷gG0JH$asnc1!CTN7 Jlm.!B1P.u>M!ȍhaF#`\Q2̤W$n6$qZ1I:jGVZpJ B 3$h|hۈFDOYٻFn#%T»R_s9Ijm54HyPeFǦzo=5~&ZS~t xKj D%"רKn l(a nv#CU=is˗F6*ebL+kH '2e :2G-֘G *pezV%;cFe`8r>R$U*0 9\]$@SwLR  ay jw46 왵cP85Q Oz'dLm[:S4_z t%kq'U:es&jj㢥,jpu8 R<`:P j>hN'4(\2Deiڡ-J,Y\D!?yҐZCT`8~ +St蹭 Rz (qr?D'fPx;E+F.gww*-TQ/Tfnc(D #Q,HJ 3:PvR0 U^̣x}t>(np;DQtZ8%Lz- $'QZAoN)!(uiaA+IaRÜ+)+(ʒ D2xކ`mPDƼ(`fڴPpj`sϡD`WfߩpTXq*n!(qFJ0w*\b#nΌA&3LީT - @&Qa@q[(Aѳ8z%vhBwLG;Ŏ~4a4NR%',+&6@we3G]|-v4She/m@GԢ_L5tV@ 7ߤk*نW6JU6g,QniKmQ*7RX R8ȎPh蕌f20"kՔx!5isPl d@d@S@30R9eEIO+eT0@k*+ooKCN05A`;p[@DTdnϴЖx=Ϸt#c<8}>=?oӤ>sQBZZ)N--Z)}q! |pC p`v6Q q\ е@@G.Nl_3oyIgJBj{NUTRY-vzZM) WA<fHϤ9hA@t?Jr8$w9Hvͱ+VWl)cjT~J[8#b9P_0[m?L/Vm.FwfbrVNp&h 9x' :Ǎ Gʀ@סXycT XAN5R?\nMȷwUiYdBIn._^.҂ W7i|H@a' ړZ*v.ձR j& =Ew5[Oh )n~u&zL4j b0_ϯGJ&sV`C"BDbϊ]~X;jwLd+k7u}$#JW() c(B- 8|R||qo!i 썲Q?N)Z)`cyKA& .)hqc?b&xwzɣup*+mo{X 5uZCNjno{'X{cmmJTV{HM:VA>1B: >9r#w^S3Ő޺e@:bRaL:;ʽKTf}:]JJݶ2Uuvn`Z@ ncdn]]7;(u`u'%Ogǧ%SԋLuΦ @nM+&/;d$L/!t[9tb13dHbR{*IUګK(5J56())cЎ6J( RN'(D^u]*@JŴ|K:LXZa]3̷h9%cp_zrgJxnKۜf=&AҔepB³fx}cxvxqA9Y6P¨^K*6ݵ v=[ ?TKON|'\FFYaƂ [ͪՙɈؒ%YFqZiWO)[. 68p)ˣi/}wm xL%'N)oVppؓ}Ur#AsԫWG)ϼK ,qQ>eK1!(TlRܡ{W9$ߍ:̂`hFC7$0 Ĕ(yYapIXPA+Bo#٭V ϓ4#&mm9k=sR&ۧ;vt]0^(SI±j٪*29(N6D-}=6N4!zLYH*^ATQ}vd' L8](t)GzJ%ZTӠd%ILaX^7c zl3ʐ^+#u3#0%pd#*zk3GAi(B*UI2BcA%Ԫs: jF>[ʹBCqh9{4;%e!xG0 jZ)Db9sR/)75;'Gt(3Pf=%Pʣˑ#F{c9r\{YAf Cy-@nؔvF`ӫxoW!qHY0@zKDŽg+5KL?PBbgW +"iٯr4XOK`DQ ~ WUAMKQ9\Gym}h༫} iIqݓblMFG=)qt%؉18[ktW+[GM{LR N&ʚSxq?O:f1S# MtM`e1w>˼4G02!ŘHz2^\H5Fj Iώ2zCd^_D#5B'2~37P&aMJ} !0 i6%K}Sm%[Ko-YphYY@0 "%\EGSG0ν}ZI[Y+W(n O5zJTKGwו^g>r qvw3Zb>6[/,=w[G s?ikj0<? %lERۦꃶch_ڲh75j3=okJZL 4h4: qIHy+L_U'%Yzc5ˋgZF?:f]\u㿢X!Hw5[R3`Am9nP>"upõ*LpVKUu *3FY:Ȁl[汭V:s[~SBף_7)=cqN ١i\wl Y漺:d2}5:[!H1u8iÙ=@ 5ytI.ݨ+vrGhk=\c tCIpB0J$JgxN}$/%68Bɕƒ 7gEHJ %qh [b Jel "PmL7,աQ@ \Z(Za9hzw]_,Pzܢrk*d۪.ӾaD돬_¸㝆+k[LbcNI=ɿM0b2*V[ Fl `7 ,+*%DyvgOᇺ۶)5wgh7@ R Vh^k\j }(1igUn~E&$" 84L[K''(Z]\U y >Jϣ OiYj|wiUmҗ%G*3!=u7U:(%Ƙ(c)-sH gRsdtr?׌ Q{v~%5bir0D7tn4u.gfrRW̩1Rja%s{{Hn#K~؝&Afy-{$=`J]YUbU[peExn fK`m*b5kZʧy 3 yP ,naA\iU {kZO\<BqFO,]pxcQYI*F҇1GcB Xp1`5xc.%ߩ] * d nF/kdmV[`'ƅ/s4r|5ӎ2ݝQ[Q0- #LH9h4n4(DP\rR`v2ͳ2kTZ'< N"$Eʥ*b%jKRI.BS'QZ<0AU""mdk@pLDEL/*MW+9P+ 1SNrllW#;Vhja7:dtP cp{CCH%E ט{$2E,q#l̷0ƥV~]9,TyO ՠ{Ƞ6 1PDȴA1: :!G _<;K?@s2?YiEkVPBF設H!O,٠(&qX~,͒|w1"a ƪ4DX;)&*䨣*Fq蒜|-8P> ΟGZ#usrz23$ДҜT5_t 1n)8A"Knqa3A^XIgGQ(#3A8$FAw$0 K@ $1FVZ9ί)◗dlR3Y!ovQ).W;=HmMh]3ʹ@i۩>t\y> /7_fYk vz]mI`m{+/&WfBBL!3JLX+꾣 wǨ{;뀏 M3]u{N%aT.$wHLcȮ xc6$xQQW6]JŝS !p ;%uXavJy!ޑS<$GMMG .։TuHJ; 2"BIA>؄Z I-{&@=g9/]ZWV<;YH+ & ]a)qv+RSNjv2ͅѽXlCp=¶->RrN#k&mZJ*g1ֵJ5>T+ij$xqZ1|R<Rx/iτqYI{'0mo i&SV8WC 1jţg9goEhNDՅZ(/oڠuN޻~?zvd=w{F}܎zUHo:5fkn}#o>ྶӸop1y*T ){KϾοQxl5k]bOX^hXX.#lZXAi:Y4lo=Ayu 02Ę@v_/}eq58ktRsʝRsqlp.m]5fuw;sӾpVӶ}7̱^^ZoO/l/޻MȾIUɈӾS;3t϶7B.辋vVNUyw|ܵ;|^$ӢŝZ^y7 vw*95HsZ(Q?[b7řUϐR>aq1 r?o:Lr2Tj:hQ >ԓ(eV;lՋ~vyðF*JMe2ر׊t4AYAϔ!#԰Ou8` jg/g<+UeKP׹NBVIw4izUlo5ڎCYy~5nR4: U$CE Cr`CU k eK|?NjrP@8w.P=0c"^FWNvz/De꜏~5;2\d sQ"뗍{|rèȴ:0zz7p_i[Dh _oO}B"/z4Cg#_$ kdFtf:oLmX. ^~o, D'F_ݍs.F!_'4Z) ~9]qlf0M{` _(I$o7E30vd`Y(:`KI--&Jc(jFOi` }0Uq2L@Fu;4iaOC ſ-}w Xʉ髛UnnG*t2OƣVC}Z`/ yXz[̟?c +Ș;B 9VHq0Z꣦Z-W%]$q}_?.~JL]J+y̭V-|oiI]Zac,w㈠9إ'z] uyKs22 D J:}A B  ͆G~S&*뗐\њwTG& }2v _3UTwJOTflr*SPh AT=J{#p:H#$1FVZH:-Y0v3YdSVPwFmE yb E2j ZqAŬ0`F w sOD&#1 [`|Yc:J)<k\\y+7e$<\ n o(f"sU𲐠€f2 1,u\AJՅhPN&֬!@Dg! b)! *yT h, |($& )U` %8(? \z42kn,s P&Z P֙(cz6_!r+y~bcM4;1K֖]g %۴d'ň@Yˬ Xy] 9Fx ϙ[ 0Z!X#8Lh %$ h-DܗBdpIډ__7%;7i_g_ /٭JÏ]}48u.]rE6( Miۼ3AN1̾ٵHܱ)aTXfk@*a0< bbL" ν5&eRܛ]kiܛ]'i)րotCy&ՖIhI`3xd?Ђp\&m5-FŀŌYd~1ShbV_ +Ŝ9UGF׸fb.INbQzձłȣ;^X q~K$/xŒ%o_Lr3NV "GʝƠ"+ (X8/uR6;k/Κefb/Vxbh}YX5/ר_GkD/H~StZ3#ig,]Y芳IXN52Q1ŏ^Yq\YLl d)v d5j]:n>¸/?Y%¥V';;S֥VT'SMߍdtw9tb]ʝ&ÊmNuwqwd7nwjG Iuv.JTt<"_yn[Lt=#37YSmpa~zyHDVcLXv4E~TڮؠZ(uC0_gh>Gi<>wu4$.O;םβK9?/\6N&7 yݞ!U\HcIyi|Kl~0N4TW)I`49boRꑐ#$ &E)UߠuKŠ褎uk.˖Om M[r"$Sr&6b&6aR1":clcݚKEV):dK[r""SOH[7`R1":clcݚK2i?7,RѺ!!G.I2.6[*UD'um[s(zo-uCB\Dd l8u#AIc\x|ϊ֭ 9rM)búQNºbP 8V5ƚ["On-Uk5HȑhL)z S 9T NXYM[r"$SﱱrӺ)!"2MĠ۟8VisͲS[i r""Szf= oX7 DU!J)1*4W93E1K Ct5\&h%[Wc1CdJ՘%j_YK%C ~8&f:+߃+wgfzO@g'BٻaW~q~rIsaJ-}f+$M-!H ju ^} RTnշ83(}NjVE+HהYC_SH{ɚȝq1L{l1!w` s$ .(s9.ykM]Py`EΟ $[W$exid' ~l^wQ?>? ~J0'KQooLϭ:d$NO*2rGIn)gzD1jdpgΥ b( ڰu! #P2 fq;!0 e ,Q a5Dsr ĸtJk!5qfSR /8\uǸ'X;xJB.@ \ !`WR-J4+{~rÌP*!,'BB$"{ *͙oL\9a S\Q*}ˉ,$d +9Z0! JG=qPoꈗy6WFAHWnxŏ~^{zrWB*{^t6_͗Y>_N+3ߤͪ~ghG-ʒBh*ݷpN^Uyϔ"z&?-;Ef; jF` ј@5"te9mL0?(X\JU͍+`(G V3p2ʄ 0"hj@9x߀<Lŏ^f2,ru>2֟L{g]ŚwV/FR@Ŋ^3[pr| *SB\ ސQ39|!.u|_X==9o$Tm_>[p9l"/C,-E;Z%\ õ!5 =[Y /ߒ{g;`ɝ-ë8+L_63#-f~Y'd?VxL#1n<+ESja  kP `.v9^a*W":VқN2ޏJGףte.p|'O}7Tӟ8$^-L8zJU ê}`?uJct= w;WQ8pb2qޛ4@$ɀ4=)8_+e;CƹӬU9+E,\VBE6("d#jLCP#m&;e(F#Mhn*R*F*\ (g4\ݩ<ṯv2b)F2*< 9% w& 9D8 tLŹ2 3CsJQ AeoO2H.P^.WeRDʛg1ɻVm{.םjeNVcb]ᇛ&ǫϷa,DkrrF m r6"p,&VHOZ|3%EL,.ȼ] ָUW6L{ ),R z>ye_~aI͆oGE`I)b 6TJbMd)lDcv0,(0E5Z&q):Oe΀gSLpLQo]n~nꪓ0SM*E e 6 Mo|x^Cr썥}3D?Xh1ЛI"JዊAj0'gzLi,\%^?2F ˜qY{SՃJi~H_NR݂͟zH|V䋝/Ȍ 1{go=̽KAk`g_\*]ZG!Dҿf%)+uҰ}?Q[BW?b) +*|qfv')9~Ї?o2*eT˨ eeX=N㾳cm@B,i*gb"sEy - {ǹg`)4[Čg߻:ijy@!ClĨ:GpίoפUy֛b.k{Smu g0/;@ O94153b.HMlQO&U;UBKNɘ4+FzD[T[ xlMi6!c[v3XoHFfuR>N\?н$Tn|Ҝ Rt>ljY~Ŭ͏m1[[β빻=syY05܆L .L6u OsqnіGdp؁yǷOײq\':g O+F/ eA9v`kϾtYH*UW m$F|Gz!EAWJ^=u\zlECns}bѲFuF-p;ucYߺ͞b-?Σu_e923@Rac4%NZH<,j! 6rHVu^ ՘WL6d'lcX$A _ bBI J j\Y {}[le߿{A~56{F[>tdu>LYA23&r#oyy$3d%5fɦLԀMr*>KApsL| b 0{-LRJ*U=^S4Ҏa7ؾ%a]Et׾ʷFd:i9aCg&u`w4J9c pŖ\YNr_lK"jcKF(/kݝYȥdR[q%."nwkXԝ92VjC3ua`v ߙ߶+o 'R#$< ts<,v1] 1c#ַEdI)@`-:(pjé9wz:eFd #c: FBnr)p†۞D%ߺB|5wz՘y<0V 9%&(NcL|b,-h‹쬊yEilO3cRv&Zi;5vgb>O'#@"E)Wq~@}5I)48ec>;± \$tݛoGe5\|GUy["a!(@@i隄r}x|H ye'lքǛքw|,@ {0J,&u;FN %mn2ٷ=mz>qo^h3SKW;~Hq{UkWQ 5梀sq#} B8$| mi۬GinCw셆{qY mO 9W$k)|xwqN[ȴr~څ*Q9/. @APro;|:'=~+73|^O!.;3?˂T O$*;5ZЗ>^ rYG>Mj:܆6$YMp_/tû7ͥoͶ] $\ |TaImLbc5k}HlZl\/çt:ý%X-׌Xo z#hJFPm,dL=c<\BsQW2܋OނLm}<~t _bޱ:$םF nwXuhݬ=Z<9/~J +(bWCm%h3mb87q+s =<2\7GӐE2Cm]si`JQ혾a_4}Sn[6Y?l64ɿۧ^5fVE6I9,t2-E'KWS՜E_DnߖY4^Ww*l0Քx}Prn= :ݭ].۟BY%4 ɢ5^3:ze0fW2q #v}`qjyxC/w]RQk륰l庫 'l2PXicf K3`LFImCt@yƺÏk层xl,Wm'^vf:ݢU:(?BղsbY T)\3v~㖯q$*rV, 1+GR(8v"u+=ǫ{ru9T-"_5uJk!8W#/͙#hu\ 1a^!e`9k24- uQ o|ʎlԹ+r?2:@_6(!"f5u) -L9I"(ٔ?u{I?3#Ckf=J 8QjWp-8kiZt0<6eM&ebfw6NOiiͰSs*XiYKPRǘq}#H:LNc ^6õU% W100)/TNcF!dh D (.ǔkdT :Bg% p0ƒC̅шN{a %LtTXha(6;yheg vllv*1Beڈ~Z/6C* %)tCtJe?Tq'F1V?K{FN=ݟuR.|4Fq& k1)# `w ZL:EqԩvSGu1w$[wZ) Xi׻SIJ_D8zA.yQ2HD:,DFzm3bQ J @DÔCTK\F\ wv9%9E3yNhafNE$e31 mz&BmECz,U [!Ʌ[8,2G,1&A˒S 4P% L瀄5'ik!fH#^3!rR2S.G~bTl(Y2C44xgx9٬E mD @[nT!'ʁֹI*LuMĽғ7A I`Gm :eE{M{gEr9h̞^!J|{Be׽x k$Q2.&fq cxO~ Rk%7&[?оVdB֝#Yd+COwol@ ThN JFȄran[,fP??@@Ӊ{>ܞGDu=&kE}SǕ}Yu2.ل;&HF P#h\EN{X2P .%RA5Bu w^߾6_4w/noId"5HgoNNs#l'mn$z''e1#*2jC hu-JLwَEo.At1~Im?{UnT:0seapFv5FSQAa0c#*DdY&9ItZFu$r!&N#E7w7.) fTkWm 5'Z)!m tsO.Ȏ}Ogg;mt3Vם:NO$.8lxsy/vkB{J~ܦsSݥ|Gsxe4$PӱF%Y_65~}W5hnҖ=F-&wG|VWO)fAhF놣ީf2L;A;/5ߤD+g26clLYm;[=f kC%ء ^EWbZ*iJ~GqwoM@L7>ǃX|sܨ܉1Ѝger>D9Ǥ_ܿ8+d?DЏzR.؜QEISs{םIgns!.A1tܻu⧰''\%]zH!ߩ?tz7e6AS^|\[h՛o!cٻFn$rX,Zf,>${8 ,6ȗ[|f[ޙ+_ʰݒG$-՛dUY(x#g 6,ň.\΋r9/(e-9#ND2Wʹu`*ŵ3Z9tH81mYN4-W rX{p֋HteɡkArDg3z}T (&P,vL܇VжVoYs'V& F|H+dtwăIqNw[lO !r:=XBOŬ,ZP#Nƀ0IN)*z4BCD?t:QW%'ﳓr vsww}ߝe>PH>{݌rVZmw:VYr8e$zrF^ X q>j~EjyFhtP=dIQzၑ `mIyN(V \ WVթ "PZrP O-W/a@B1rF1Ԋ&/:llghRreߜ%Wc|ݥZ6xYHk-?~Ӹg|.ZC<ۧO(2-k\>DxsvW rח?23_9G%t]0#X۳g4 V%.C^^.z1y8Z*d)z8(MMWFi8]?/WW4TVp/A#![NCICTCOPV~YOuZGf'? 0!mi!4+Z!;JOS04B R~OSXg4[*9 7(A%CcǍLH{aO΄@Ҹ΄T1D&XJXg4ȫÄ9#+/:8FQrm䊡=ZRWDp.|YvGzƙA8#/_jbj/i|Rf:$%%|Lۮ1|kRkt%cлjū](LDJե0 CV;C yz}S;tl&3iɨKUɗ1l 5fx B2фҷG2bB)t*tdf> E\dIi$_ķ"n -` {z9_ЂrGqՌR9k-ΥQc K;vu85 8Y~rDڎ12k2\vjOYMxF16itfȿiO(J~k2D{r~K6B`ɕՓ-5ئuW+_~Ⱦ4vݗ#gU?[٧'Pdt#ƧM g{9aK[yr`:.%Xම;oCWێxPVn;XcG15M@(9c#ߛ?5^ }pԝpQIT )smxߤ;>E>+LݾTPL˪W+,]-_@`O} ?8hUƅM&調MrTr/%[3d骟؆Dՠ+ŭ&_QՆɩkү.^ \.7dWռxnbq*J~Y33qͭNm]:]aʢc mcXY6eS?ʓj/-򨞑}B჉nyÀLC撚WLb^oÏ"I4ګ+zV=6^d| IcpkT}*MSÛQ7m}ڕZ)vzCybd> X>jvR͹>A (@xꓻK`-ldG? t\0uOxH}&OgB/>_o/{_ EP{Dԩϕ=[I|0?Ejt2ؕV]wG3x92tGcO5qr#j pgvx^9\z*}b~$r;؀]v%R֓ j, j[ex2n(,16ij9濇!v\Њ}xvp);omݢ@YE@Vߓ p>g~~m}'G^RKuvCW.S]4}zf I.w󗍿%eAC_>U>/UKRE|⺱i.u2d`yYIl=Wpc9rf8vS͏-D}w{K/ϳj6tw3){qوm9[OK~X]os֯uҴD}>Xˑ5,[WYEpyMhQ25t1Faǚo R21'w *"Vh`!zLN0!ZY&-Emzѳ^iXckhXjGڃ65`b@x<מKAbTS[+ZA4|^4&sM2UCȌCa2jSX`ݨl7,Fv] 9}א()36ʩ WJ Gq+Jcd)jx & dQ63?H!mǐĔ 6=N}9q')cR iҤY &YǴfH?VY9c.&K\(0B {(B*ҷK,FC̪kVʱu$ז*hPZMo־wH5(XQ*91z8r/yp%Y~^x$eJ^׮MBo}_EE=pyAG_wN"$LNMw/IQWyKZ}-??ں$.}4Z5_u{X MNlti}^+GW.I@'tBo^+% cv }_KP , ԮQIyF:6&K#2/UaV`guCE'7wY\HW#C:b&)|GJ"6\FNX(OB NNdo7:ӕsc@r&|p]w@.T8Nw O^=/z^x|W36E$Ri0>)g^,l&˹zDN͜ GSL:/]ê)CKj(|GK\W #X3rk87BP ɎQ2n KVW}5RE~>[y8%,-V Qu?%C Alàtd7~}l i]dHb,#73HQ@llFpmJ+Pm"RxP3YUx0t^?ǻۑa0LW8(Z$с:fi)-%mL.\jƭwmmPiA$*^s2,#=k^)P{ZpI޷`nJW޼Dێ1c$zVV@>W,iHQҲQ>~^{/[a,S{^/_&}gtg#20_CDbȊ\ccy&Lyh1Wt"[3IՉxoJwKߨ_Fi=1ٛ#4+uWf5y݂@a;9/^1W~ڢAGP>[[e d.tihw^'=1h2蔔KDg`)ެf.QE#  X)99+[fC23 6rVxp&PJ6;H3FwڗVF*gmѧдi_4~.VD?{F/P5vy9ޤNr\ٓMpmI^4HIRpHJvKu@ߔrӼtE*Z+D-Nwi3>PZ^P\,E^a#R1 dcRTql{p qxPEslE5B`dRHd% sP׈Ʈx:4z 5z+t|^(zWFY$!}.^w@킛lPܸߟn/gg -'D&4hzx0R::$T?UYsEGYX'cdk!YCd a&̡"F)A&jo؊?{7wj55[G OG_Bц3S/-y@neϦrƄ;:sD:ל=PK,ئS !}[8f0;ZKJ^SҼ}";V~?#BPF=o.zeFOtW@}E}t=T%܄gv>Ɯm2!q&ҫ$9q1ɳ=AmUt&V%G1ҦXKnlə'㢷RB1J+!ҹMItZp接pf(-ކsYT%tWl7+̕mSG^p}%˕vϗ3 [Cٓіj\@xɍ_}\m"ۛ4f [;+U ﺇqte:֑1:p~{zv9f*ot4-=9GIuxNv[+1..hpZJx|G/o~FgK7q18{AW7 ?L&dٓwv}&v SC]N]C'<]?:!w!&_.xs~?,H̔djU|\9^]?N^x6p#thtאq)P#7%7ǟ߷>.׌p|%+Edz-.n'7)͉szOe>}jo;'˷哅9:f1WFsJ'bO?HL&gj s!z_KP+C?^'@F{̠ ē]Jwh h-%j-rڴϢZ:]뛯 Tmp5^zl͊ɬЬz9pCw+6.e*ye*@UEWTqD3-ȫ `.ʃn<ֶ40a߱f;0b dwh͌د}[Obx5rXFaQ@)(ti*AʂVgPΚi|++=RPf/D$5*LF.<ɕ8]AH|dq)^ :Ƃ,im Y^[nBfI FD[ ,JJ-[T*Mc֙6fZew*)TFăResa%4r+qcPSy]{.jmrx1R\##ǜarLU:w-g;(YL 3"f#%- IˊlI. SYX;BeUs#FJlލ?k^NQgdPߧMMnNݑXnX=eT5 3[Ls%찃saYԟDL6h|hQ4;c ;,ָ̚:"Ԍ0i[W%Л-3/E(x1|\خbԚwvq-]jjء.e_U&PwxWy|a* ;U?u2J9i񪗺DN{ =CA*U#n@ZrcȢ1U,Y+ʫ$YAeEqJ RƺB [j԰l {G 5G,1|0"rҥmא2T&yo13"xtU^j+aԴ+IѺ jNe47'Q!2'a)-&m)8)  e X\#Srm_j ^ [oeݥ P6DQFN ѲҧfbpE V:`W{ QؚΌ0_ueɴSFʑ&y.=tGpoQ)X]Ag|D*-K@K٘bjL0"%+:D)~t7- \N&wX[3'%`}h[9hybr6q%)l+ӪKXU ucA iUlY{lcݟN]vhVâ5xތBš9o%mx]s {8k'AࡸlSܿ%ȽD]X){2t{|\^} 5^ڕn<.GBb;HE|4-]{f69pqmFڧ2s 2#㘂R#O$"aK5ZjZS ^a)1 E0egQlx1+9+s#ZVhY 2rk * Ѝxe`uˆVp߅k+3tMs5 8 X4$ZR&3w DE/ CgܣeCDC)P9 )0`6ެ[w@+7z-n+C) 4z JxjDvK-EE1\7 -ఁc֡&%X!r!T.zkbٲH5(V@+MA`ą1$&xWfF vUc1H%,V'vˁƞA<٬L8Mn_.[TlَNFNLցV!ν.{{S5{@C~zc6"6F!'~8hZfh4SsāR.DxQ) + ABƏW w5YL ػ\~zYU st*Mp :+S Ryܮ)Uye~{7 y5ѬS~Lپ;ߚut|~܆ .Sy0T߼m.{vy'^/J)$QnwV%jȗM>)>kԥNe-HZK]R_;/[ 7dصnw廭(bVHHB;-q64 H\St͉\'sgcl(›F-ׄ`+&RNJftZb9Db;K4C\Pઇw߅[Ȱu/bY7׳d=F$F)VmoyFĴ%E(B9O"jD}I']^(YX#z)6S_=Z}>mJwJ6Ehz0;>I,`?KKF*B1}@~fvf;<зК2jP?r:ɺc扡BF7G{;?\~rZ(vǝvKc/PVaǠtQ q!;*.Ѵ; d?L}Badg!זo_o `%BpcuHKb!3Tj/M [ -s4+]I3#chzM{1}'WU\'2MF+}GǥPļr FQXu@@!0 V#WP[M7ڳh$WijI9hz{X =Z#o?O?&NVQ{2S)a1p^NTovWU=x~Uݹ{8:p+R\ WjRr)S@hR̗sBzN*瘯H)8i^*]UȯxO2(?x~U (^eBqJZf"]3n"Vs3*tF,P|sȒ7mوy!rD]ݛ2%gy<`VPQTdT# N׽vF)Vrdw&@/'J' *ZO+# C"$0v̋ٞX_Э^\sN>(TNe¤l*J UC{K.q2!L/ob^Y 6^nkWxEg6#=* v%d0gѨ,qnz{7=;kyq:2 }/N+O?F{5K{5_\"&F~PRFY%݄=;oQ̏{֌[ sy|{3WH%az:ri8iM;o^8;dNv!eYY7bPzg֛!$[|F)Hf8֚j&_uQhfQIAsǠ@j11NTC(pn|D4א˛SSZ0U6\r#+PU/:?7@T휿K'(*E#g(y Iگu $hKךDA{VJ0 RBe5%*PQ%g"s%ԦNPM!нNL q\]LS66J:wr/f; k9z3+TBin KUٟ?yO翱Q.s]Xq4…F!R֮F_#Tu].[3cՃ4+fŷ^ԁ%Y9|e;%l9olc( U`͍u@`])ii2HQV\E!@DEɃX.x?U٬WSIWK ?_/rQֆV]d&j]2{ w^+?xi|Ԣ\Qc.׹%=ټBPt^s~dսCF$:mڬ=jg6z Ƕޒ[KowQMBIWb0DNyB" ;OJü& uUvaJ; !HZęԂ:Ќ) {[ rs ԐVu.Z H&T%:P \-[5<5["\6k)Ir1 Bd$%olʠeZrcH*sn@+6 lSe84^ 7` `-~\@o ƀa) V;;%5[ ZtͿ-^NUr|Z"j-yAwh+C]v~3r]k{۵Ο^k-3vEkKf5\zoIFqyt\Cs|WA[jWsѝj&Eugpdb a1C&)Md~ N&:]B7܀- |TsAʌsUgE{f]ՙ,L nŖN?/}<[=xVR>Gp?8Wؾ]o7H7gb;Ak+>O׵}~-b;2`={A폁}I13]U?Z[YeF=ZpMEۻϰB ݍ=M,9T{]iPIMQp#\#’w˼eԼ[*ٞoDDPDD?NO3ނ$7tEZqqK[Wʦ^ԯl@ aI0̪a>PG˱J~?ڈr!Ir8چN2IR^J=Di ^er"\ (8Cj&Ȩ,'&CYF"Q9B+gzȖSӋ{)ٿ3Ub$#v #,a~Vd~neD#fs'4 jT RbXx!*PjQ~06Q欲AhF=4zĵF1GDK #|Gsy ӱ>-2ffL,OcBfEq@!BMۻ+{ Eʏ#.z0rV>9=Y"zb0اJ y݉pX:,(&ìG{3m<Mѕ b{]؝h`\oȬ)WP3qdpe'BM!n$}@p}^4'2c VL?y-A d5M})(J0b\uXcC@]L4ZZ%;uT)PE(i!$&:F9jI#MOas..Ev6" o9P%KIhZ'%3k5RۘHIּ{2>ܨm`Ҡ0`4m^j8HfuEqٗ6 `\NMd!rHZȯ]xۀK=6bIJHӔ֐PpڃX܈axL5..m0'F $j6n&~x v ÷4iphƔ;)h^Vں*7rEjSLU-gc!KO,,0MIKȆ27k{R|$8zR\OFOG $ Hp'ƹ_3EwjFWbaw7}edCz}'57aUr1|"rbXˁqkl9=NQݹ=ՆVFeݮ9vG'lE(8o}"Dٴ g4,ĉUߞ]wM?;cssf~嗇Q̾&{2=]o̊>lntf\i6zz,Q_iM3V;_֔oIh7)<3h=z(S%wݺ6j>V;_]֔$=' (чݺbPt~F(;nj>V;_ה6T:+c_bK7;mx_XJHԞ 6k5>y!RAKް!Tf iyf5RB)5Ш"jZ#轿"xϪ1Ds/{(=I+캲m;ԪTJP2'dT-rӀmxRhcZ l46j*$SV,7=|B_$oWh53o*\ZO̫ږb[FJr ǂ+1'#A,+;ыs}ǩP*q`R nȜZFn!\ߦܼ0Nyf<~yO sd&5sB)4扶 AVVB6K*7?}a<34VZdvar/home/core/zuul-output/logs/kubelet.log0000644000000000000000003122624615154404340017703 0ustar rootrootMar 12 00:06:39 crc systemd[1]: Starting Kubernetes Kubelet... Mar 12 00:06:39 crc restorecon[4686]: Relabeled /var/lib/kubelet/config.json from system_u:object_r:unlabeled_t:s0 to system_u:object_r:container_var_lib_t:s0 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/device-plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/device-plugins/kubelet.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/volumes/kubernetes.io~configmap/nginx-conf/..2025_02_23_05_40_35.4114275528/nginx.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/22e96971 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/21c98286 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8/containers/networking-console-plugin/0f1869e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c15,c25 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/46889d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/5b6a5969 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/setup/6c7921f5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4804f443 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/2a46b283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/a6b5573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/4f88ee5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c225,c458 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/5a4eee4b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c963 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/d1b160f5dda77d281dd8e69ec8d817f9/containers/kube-rbac-proxy-crio/cd87c521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c215,c682 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_33_42.2574241751/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/38602af4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/1483b002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/0346718b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/d3ed4ada not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/3bb473a5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/8cd075a9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/00ab4760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/containers/router/54a21c09 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c24 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/70478888 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/43802770 not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/955a0edc not reset as customized by admin to system_u:object_r:container_file_t:s0:c176,c499 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/bca2d009 not reset as customized by admin to system_u:object_r:container_file_t:s0:c140,c1009 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/37a5e44f-9a88-4405-be8a-b645485e7312/containers/network-operator/b295f9bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c589,c726 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..2025_02_23_05_21_22.3617465230/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-binary-copy/cnibincopy.sh not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..2025_02_23_05_21_22.2050650026/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes/kubernetes.io~configmap/cni-sysctl-allowlist/allowlist.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/bc46ea27 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5731fc1b not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/egress-router-binary-copy/5e1b2a3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/943f0936 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/3f764ee4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/cni-plugins/8695e3f9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/aed7aa86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/c64d7448 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/bond-cni-plugin/0ba16bd2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/207a939f not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/54aa8cdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/routeoverride-cni/1f5fa595 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/bf9c8153 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/47fba4ea not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni-bincopy/7ae55ce9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7906a268 not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/ce43fa69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/whereabouts-cni/7fc7ea3a not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/d8c38b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c203,c924 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/9ef015fb not reset as customized by admin to system_u:object_r:container_file_t:s0:c138,c778 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/containers/kube-multus-additional-cni-plugins/b9db6a41 not reset as customized by admin to system_u:object_r:container_file_t:s0:c574,c582 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/b1733d79 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/afccd338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/network-metrics-daemon/9df0a185 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/18938cf8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c476,c820 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/7ab4eb23 not reset as customized by admin to system_u:object_r:container_file_t:s0:c272,c818 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/containers/kube-rbac-proxy/56930be6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c432,c991 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_35.630010865 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..2025_02_23_05_21_35.1088506337/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes/kubernetes.io~configmap/ovnkube-config/ovnkube.conf not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/0d8e3722 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/d22b2e76 not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/kube-rbac-proxy/e036759f not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/2734c483 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/57878fe7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/3f3c2e58 not reset as customized by admin to system_u:object_r:container_file_t:s0:c89,c211 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/375bec3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c382,c850 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/containers/ovnkube-cluster-manager/7bc41e08 not reset as customized by admin to system_u:object_r:container_file_t:s0:c440,c975 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/48c7a72d not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/4b66701f not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/containers/download-server/a5a1c202 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..2025_02_23_05_21_40.3350632666/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-cert-acceptance-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/ovnkube-identity-cm/additional-pod-admission-cond.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..2025_02_23_05_21_40.1388695756 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/volumes/kubernetes.io~configmap/env-overrides/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/26f3df5b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/6d8fb21d not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/webhook/50e94777 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208473b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/ec9e08ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3b787c39 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/208eaed5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/93aa3a2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/ef543e1b-8068-4ea3-b32a-61027b32e95d/containers/approver/3c697968 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c22 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/ba950ec9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/cb5cdb37 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/3b6479f0-333b-4a96-9adf-2099afdc2447/containers/network-check-target-container/f2df9827 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..2025_02_23_05_22_30.473230615/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_24_06_22_02.1904938450/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/fedaa673 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/9ca2df95 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/machine-config-operator/b2d7460e not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2207853c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/241c1c29 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/containers/kube-rbac-proxy/2d910eaf not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/..2025_02_23_05_23_49.3726007728/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/..2025_02_23_05_23_49.841175008/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/etcd-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178 not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.843437178/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/c6c0f2e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/399edc97 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8049f7cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/0cec5484 not reset as customized by admin to system_u:object_r:container_file_t:s0:c263,c871 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/312446d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c406,c828 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/containers/etcd-operator/8e56a35d not reset as customized by admin to system_u:object_r:container_file_t:s0:c84,c419 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.133159589/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/2d30ddb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/eca8053d not reset as customized by admin to system_u:object_r:container_file_t:s0:c380,c909 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/c3a25c9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c168,c522 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/containers/kube-controller-manager-operator/b9609c22 not reset as customized by admin to system_u:object_r:container_file_t:s0:c108,c511 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/e8b0eca9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/b36a9c3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/dns-operator/38af7b07 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/ae821620 not reset as customized by admin to system_u:object_r:container_file_t:s0:c106,c418 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/baa23338 not reset as customized by admin to system_u:object_r:container_file_t:s0:c529,c711 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/containers/kube-rbac-proxy/2c534809 not reset as customized by admin to system_u:object_r:container_file_t:s0:c968,c969 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3532625537/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/59b29eae not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/c91a8e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c381 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/4d87494a not reset as customized by admin to system_u:object_r:container_file_t:s0:c442,c857 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/containers/kube-scheduler-operator-container/1e33ca63 not reset as customized by admin to system_u:object_r:container_file_t:s0:c661,c999 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/8dea7be2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d0b04a99 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/kube-rbac-proxy/d84f01e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/4109059b not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/a7258a3e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/containers/package-server-manager/05bdf2b6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/f3261b51 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/315d045e not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/5fdcf278 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/d053f757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/containers/control-plane-machine-set-operator/c2850dc7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..2025_02_23_05_22_30.2390596521/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes/kubernetes.io~configmap/marketplace-trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fcfb0b2b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c7ac9b7d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/fa0c0d52 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/c609b6ba not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/2be6c296 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/89a32653 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/4eb9afeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/containers/marketplace-operator/13af6efa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/b03f9724 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/e3d105cc not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/containers/olm-operator/3aed4d83 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1906041176/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/0765fa6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/2cefc627 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/3dcc6345 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/containers/kube-storage-version-migrator-operator/365af391 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c18 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-SelfManagedHA-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-TechPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-DevPreviewNoUpgrade.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes/kubernetes.io~empty-dir/available-featuregates/featureGate-Hypershift-Default.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b1130c0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/236a5913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-api/b9432e26 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/5ddb0e3f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/986dc4fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/8a23ff9a not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/9728ae68 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/containers/openshift-config-operator/665f31d0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c12 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1255385357/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/..2025_02_23_05_23_57.573792656/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/service-ca-bundle/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_22_30.3254245399/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes/kubernetes.io~configmap/trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/136c9b42 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/98a1575b not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/cac69136 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/5deb77a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/containers/authentication-operator/2ae53400 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3608339744/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes/kubernetes.io~configmap/config/operator-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/e46f2326 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/dc688d3c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/3497c3cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/containers/service-ca-operator/177eb008 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.3819292994/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/af5a2afa not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/d780cb1f not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/49b0f374 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/containers/openshift-apiserver-operator/26fbb125 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.3244779536/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/cf14125a not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/b7f86972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/e51d739c not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/88ba6a69 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/669a9acf not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/5cd51231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/75349ec7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/15c26839 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/45023dcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/ingress-operator/2bb66a50 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/64d03bdd not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/ab8e7ca0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/containers/kube-rbac-proxy/bb9be25f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c11 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_22_30.2034221258/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/9a0b61d3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/d471b9d2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/containers/cluster-image-registry-operator/8cb76b8e not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/11a00840 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/ec355a92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/containers/catalog-operator/992f735e not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..2025_02_23_05_22_30.1782968797/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d59cdbbc not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/72133ff0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/c56c834c not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/d13724c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/containers/openshift-controller-manager-operator/0a498258 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c14 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa471982 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fc900d92 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/containers/machine-config-server/fa7d68da not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/4bacf9b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/424021b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/migrator/fc2e31a3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/f51eefac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/c8997f2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/containers/graceful-termination/7481f599 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c22 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..2025_02_23_05_22_49.2255460704/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes/kubernetes.io~configmap/signing-cabundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/fdafea19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/d0e1c571 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/ee398915 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/containers/service-ca-controller/682bb6b8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c22 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a3e67855 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/a989f289 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/setup/915431bd not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/7796fdab not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/dcdb5f19 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-ensure-env-vars/a3aaa88c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/5508e3e6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/160585de not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-resources-copy/e99f8da3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/8bc85570 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/a5861c91 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcdctl/84db1135 not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/9e1a6043 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/c1aba1c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd/d55ccd6d not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/971cc9f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/8f2e3dcf not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-metrics/ceb35e9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/1c192745 not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/5209e501 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-readyz/f83de4df not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/e7b978ac not reset as customized by admin to system_u:object_r:container_file_t:s0:c294,c884 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/c64304a1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c1016 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/2139d3e2895fc6797b9c76a1b4c9886d/containers/etcd-rev/5384386b not reset as customized by admin to system_u:object_r:container_file_t:s0:c666,c920 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/cce3e3ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/multus-admission-controller/8fb75465 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/740f573e not reset as customized by admin to system_u:object_r:container_file_t:s0:c435,c756 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/containers/kube-rbac-proxy/32fd1134 not reset as customized by admin to system_u:object_r:container_file_t:s0:c268,c620 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/0a861bd3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/80363026 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/containers/serve-healthcheck-canary/bfa952a8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c19,c24 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..2025_02_23_05_33_31.2122464563/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..2025_02_23_05_33_31.333075221 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/793bf43d not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/7db1bb6e not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/kube-rbac-proxy/4f6a0368 not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/c12c7d86 not reset as customized by admin to system_u:object_r:container_file_t:s0:c381,c387 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/36c4a773 not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/4c1e98ae not reset as customized by admin to system_u:object_r:container_file_t:s0:c142,c438 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/containers/machine-approver-controller/a4c8115c not reset as customized by admin to system_u:object_r:container_file_t:s0:c129,c158 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/setup/7db1802e not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver/a008a7ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-syncer/2c836bac not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-cert-regeneration-controller/0ce62299 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-insecure-readyz/945d2457 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/containers/kube-apiserver-check-endpoints/7d5c1dd8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c97,c980 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/advanced-cluster-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-broker-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq-streams-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amq7-interconnect-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-automation-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ansible-cloud-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry-3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bamoe-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/index.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/businessautomation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cephcsi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cincinnati-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-kube-descheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/compliance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/container-security-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/costmanagement-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cryostat-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datagrid/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devspaces/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devworkspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dpu-network-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eap/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/file-integrity-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-console/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fuse-online/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gatekeeper-operator-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jws-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kernel-module-management-hub/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kiali-ossm/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logic-operator-rhel8/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lvms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mcg-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mta-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mtv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-client-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-csi-addons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-multicluster-orchestrator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odf-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odr-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/bundle-v1.15.0.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/channel.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-cert-manager-operator/package.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-custom-metrics-autoscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-pipelines-operator-rh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-secondary-scheduler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-bridge-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/quay-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/recipe/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/red-hat-hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redhat-oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rh-service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhacs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhbk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhdh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhods-prometheus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhpam-kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhsso-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rook-ceph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/run-once-duration-override-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sandboxed-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/security-profiles-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/serverless-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-registry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/servicemeshoperator3/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/submariner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tang-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustee-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volsync-product/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/catalog/web-terminal/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/bc8d0691 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/6b76097a not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-utilities/34d1af30 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/312ba61c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/645d5dd1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/extract-content/16e825f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/4cf51fc9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/2a23d348 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/containers/registry-server/075dbd49 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/..2025_02_24_06_09_13.3521195566/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes/kubernetes.io~configmap/serviceca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/dd585ddd not reset as customized by admin to system_u:object_r:container_file_t:s0:c377,c642 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/17ebd0ab not reset as customized by admin to system_u:object_r:container_file_t:s0:c338,c343 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/containers/node-ca/005579f4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c842,c986 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_23_05_23_11.449897510/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_23_05_23_11.1287037894 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..2025_02_23_05_23_11.1301053334/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes/kubernetes.io~configmap/audit-policies/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/bf5f3b9c not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/af276eb7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/fix-audit-permissions/ea28e322 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/692e6683 not reset as customized by admin to system_u:object_r:container_file_t:s0:c49,c263 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/871746a7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c701 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/containers/oauth-apiserver/4eb2e958 not reset as customized by admin to system_u:object_r:container_file_t:s0:c764,c897 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..2025_02_24_06_09_06.2875086261/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/console-config/console-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_09_06.286118152/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..2025_02_24_06_09_06.3865795478/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/oauth-serving-cert/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..2025_02_24_06_09_06.584414814/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/ca9b62da not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/containers/console/0edd6fce not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c25 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.2406383837/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.openshift-global-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/config/openshift-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.1071801880/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877 not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..2025_02_24_06_20_07.2494444877/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes/kubernetes.io~configmap/proxy-ca-bundles/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/containers/controller-manager/89b4555f not reset as customized by admin to system_u:object_r:container_file_t:s0:c14,c22 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..2025_02_23_05_23_22.4071100442/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes/kubernetes.io~configmap/config-volume/Corefile not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/655fcd71 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/0d43c002 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/dns/e68efd17 not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/9acf9b65 not reset as customized by admin to system_u:object_r:container_file_t:s0:c457,c841 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/5ae3ff11 not reset as customized by admin to system_u:object_r:container_file_t:s0:c55,c1022 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/containers/kube-rbac-proxy/1e59206a not reset as customized by admin to system_u:object_r:container_file_t:s0:c466,c972 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/27af16d1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c304,c1017 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/7918e729 not reset as customized by admin to system_u:object_r:container_file_t:s0:c853,c893 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/containers/dns-node-resolver/5d976d0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c585,c981 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..2025_02_23_05_38_56.1112187283/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/config/controller-config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_23_05_38_56.2839772658/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes/kubernetes.io~configmap/trusted-ca/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/d7f55cbb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/f0812073 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/1a56cbeb not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/7fdd437e not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/containers/console-operator/cdfb5652 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c25 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..2025_02_24_06_17_29.3844392896/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/etcd-serving-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..2025_02_24_06_17_29.848549803/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..2025_02_24_06_17_29.780046231/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/audit/policy.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..2025_02_24_06_17_29.2926008347/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/image-import-ca/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..2025_02_24_06_17_29.2729721485/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes/kubernetes.io~configmap/trusted-ca-bundle/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/fix-audit-permissions/fb93119e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver/f1e8fc0e not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/containers/openshift-apiserver-check-endpoints/218511f3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c336,c787 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes/kubernetes.io~empty-dir/tmpfs/k8s-webhook-server/serving-certs not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/ca8af7b3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/72cc8a75 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/containers/packageserver/6e8a3760 not reset as customized by admin to system_u:object_r:container_file_t:s0:c12,c18 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..2025_02_23_05_27_30.557428972/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes/kubernetes.io~configmap/service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4c3455c0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/2278acb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/4b453e4f not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/containers/cluster-version-operator/3ec09bda not reset as customized by admin to system_u:object_r:container_file_t:s0:c5,c6 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..2025_02_24_06_25_03.422633132/anchors/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/trusted-ca/anchors not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..2025_02_24_06_25_03.3594477318/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/image-registry.openshift-image-registry.svc.cluster.local..5000 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~configmap/registry-certificates/default-route-openshift-image-registry.apps-crc.testing not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/edk2/cacerts.bin not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/java/cacerts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/openssl/ca-bundle.trust.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/tls-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/email-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/objsign-ca-bundle.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2ae6433e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fde84897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75680d2e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/openshift-service-serving-signer_1740288168.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/facfc4fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f5a969c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CFCA_EV_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9ef4a08a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ingress-operator_1740288202.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2f332aed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/248c8271.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d10a21f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ACCVRAIZ1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a94d09e5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c9a4d3b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40193066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd8c0d63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b936d1c6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CA_Disig_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4fd49c6c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AC_RAIZ_FNMT-RCM_SERVIDORES_SEGUROS.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b81b93f0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f9a69fa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b30d5fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ANF_Secure_Server_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:39 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b433981b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93851c9e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9282e51c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7dd1bc4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Actalis_Authentication_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/930ac5d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f47b495.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e113c810.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5931b5bc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Commercial.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2b349938.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e48193cf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/302904dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a716d4ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Networking.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/93bc0acc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/86212b19.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certigna_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b727005e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbc54cab.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f51bb24c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c28a8a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AffirmTrust_Premium_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9c8dfbd4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ccc52f49.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cb1c3204.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ce5e74ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd08c599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6d41d539.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb5fa911.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e35234b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8cb5ee0f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a7c655d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f8fc53da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Amazon_Root_CA_4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/de6d66f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d41b5e2a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/41a3f684.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1df5a75f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_2011.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e36a6752.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b872f2b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9576d26b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/228f89db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_ECC_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fb717492.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d21b73c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b1b94ef.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/595e996b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Atos_TrustedRoot_Root_CA_RSA_TLS_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b46e03d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/128f4b91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_3_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81f2d2b1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Autoridad_de_Certificacion_Firmaprofesional_CIF_A62634068.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3bde41ac.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d16a5865.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_EC-384_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0179095f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ffa7f1eb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9482e63a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4dae3dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/BJCA_Global_Root_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e359ba6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7e067d03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/95aff9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7746a63.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Baltimore_CyberTrust_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/653b494a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3ad48a91.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Buypass_Class_2_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/54657681.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/82223c44.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8de2f56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2d9dafe4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d96b65e2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee64a828.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/40547a79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5a3f0ff8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a780d93.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/34d996fb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/eed8c118.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/89c02a45.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b1159c4c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/COMODO_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d6325660.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d4c339cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8312c4c1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certainly_Root_E1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8508e720.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5fdd185d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48bec511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/69105f4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0b9bc432.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Certum_Trusted_Network_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/32888f65.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b03dec0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/219d9499.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_ECC_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5acf816d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbf06781.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-01.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc99f41e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/CommScope_Public_Trust_RSA_Root-02.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/AAA_Certificate_Services.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/985c1f52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8794b4e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_BR_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e7c037b4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ef954a4e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_EV_Root_CA_1_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2add47b6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/90c5a3c8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0f3e76e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/53a1b57a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/D-TRUST_Root_Class_3_CA_2_EV_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5ad8a5d6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/68dd7389.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d04f354.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d6437c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/062cdee6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bd43e1dd.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Assured_ID_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7f3d5d1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c491639e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3513523f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/399e7759.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/feffd413.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d18e9066.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/607986c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c90bc37d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1b0f7e5c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e08bfd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Global_Root_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dd8e9d41.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed39abd0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a3418fda.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bc3f2570.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_High_Assurance_EV_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/244b5494.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/81b9768f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4be590e0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_ECC_P384_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9846683b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/252252d2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e8e7201.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_TLS_RSA4096_Root_G5.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d52c538d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c44cc0c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/DigiCert_Trusted_Root_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/75d1b2ed.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a2c66da8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ecccd8db.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust.net_Certification_Authority__2048_.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/aee5f10d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3e7271e8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0e59380.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4c3982f2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b99d060.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf64f35b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0a775a30.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/002c0b4f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cc450945.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_EC1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/106f3e4d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b3fb433b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GlobalSign.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4042bcee.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/02265526.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/455f1b52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0d69c7e1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9f727ac7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Entrust_Root_Certification_Authority_-_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5e98733a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0cd152c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dc4d6a89.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6187b673.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/FIRMAPROFESIONAL_CA_ROOT-A_WEB.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ba8887ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/068570d1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f081611a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/48a195d8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GDCA_TrustAUTH_R5_ROOT.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f6fa695.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab59055e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b92fd57f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GLOBALTRUST_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fa5da96b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ec40989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7719f463.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/GTS_Root_R1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1001acf7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f013ecaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/626dceaf.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c559d742.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1d3472b9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9479c8c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a81e292b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4bfab552.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_E46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Go_Daddy_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e071171e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/57bcb2da.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_ECC_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ab5346f4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5046c355.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HARICA_TLS_RSA_Root_CA_2021.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/865fbdf9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da0cfd1d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/85cde254.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_ECC_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cbb3f32b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureSign_RootCA11.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hellenic_Academic_and_Research_Institutions_RootCA_2015.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5860aaa6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/31188b5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/HiPKI_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c7f1359b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f15c80c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Hongkong_Post_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/09789157.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ISRG_Root_X2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/18856ac4.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e09d511.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Commercial_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cf701eeb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d06393bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/IdenTrust_Public_Sector_Root_CA_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/10531352.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Izenpe.com.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SecureTrust_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b0ed035a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsec_e-Szigno_Root_CA_2009.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8160b96c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e8651083.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2c63f966.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_ECC_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d89cda1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/01419da9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_RSA_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7a5b843.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Microsoft_RSA_Root_Certificate_Authority_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bf53fb88.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9591a472.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3afde786.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Gold_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NAVER_Global_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3fb36b73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d39b0a2c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a89d74c2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/cd58d51e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b7db1890.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/NetLock_Arany__Class_Gold__F__tan__s__tv__ny.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/988a38cb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/60afe812.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f39fc864.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5443e9e3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GB_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e73d606e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dfc0fe80.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b66938e9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1e1eab7c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/OISTE_WISeKey_Global_Root_GC_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/773e07ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c899c73.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d59297b8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ddcda989.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_1_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/749e9e03.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/52b525c7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_RootCA3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d7e8dc79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a819ef2.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/08063a00.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6b483515.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_2_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/064e0aa9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1f58a078.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6f7454b3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7fa05551.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76faf6c0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9339512a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f387163d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee37c333.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/QuoVadis_Root_CA_3_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e18bfb83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e442e424.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fe8a2cd8.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/23f4c490.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5cd81ad7.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f0c70a8d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7892ad52.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SZAFIR_ROOT_CA2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4f316efb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_EV_Root_Certification_Authority_RSA_R2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/06dc52d5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/583d0756.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Sectigo_Public_Server_Authentication_Root_R46.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_ECC.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0bf05006.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/88950faa.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9046744a.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/3c860d51.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_Root_Certification_Authority_RSA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/6fa5da56.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/33ee480d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Secure_Global_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/63a2c897.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SSL.com_TLS_ECC_Root_CA_2022.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/bdacca6f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ff34af3f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/dbff3a01.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Security_Communication_ECC_RootCA1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_C1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Class_2_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/406c9bb1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_C3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Starfield_Services_Root_Certificate_Authority_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/SwissSign_Silver_CA_-_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/99e1b953.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/T-TeleSec_GlobalRoot_Class_3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/14bc7599.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TUBITAK_Kamu_SM_SSL_Kok_Sertifikasi_-_Surum_1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Global_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/7a3adc42.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TWCA_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f459871d.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_ECC_Root_2020.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_Root_CA_-_G1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telekom_Security_TLS_RSA_Root_2023.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TeliaSonera_Root_CA_v1.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Telia_Root_CA_v2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8f103249.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f058632f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-certificates.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9bf03295.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/98aaf404.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TrustAsia_Global_Root_CA_G4.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1cef98f5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/073bfcc5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/2923b3f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f249de83.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/edcbddb5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/emSign_ECC_Root_CA_-_G3.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P256_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9b5697b0.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/1ae85e5e.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/b74d2bd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/Trustwave_Global_ECC_P384_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/d887a5bb.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9aef356c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/TunTrust_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fd64f3fc.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e13665f9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Extended_Validation_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/0f5dc4f3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/da7377f6.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/UCA_Global_G2_Root.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/c01eb047.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/304d27c3.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ed858448.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_ECC_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/f30dd6ad.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/04f60c28.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/vTrus_ECC_Root_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/USERTrust_RSA_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/fc5a8f99.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/35105088.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ee532fd5.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/XRamp_Global_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/706f604c.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/76579174.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/8d86cdd1.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/882de061.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/certSIGN_ROOT_CA_G2.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/5f618aec.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/a9d40e02.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e-Szigno_Root_CA_2017.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/e868b802.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/83e9984f.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ePKI_Root_Certification_Authority.pem not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/ca6e4ad9.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/9d6523ce.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/4b718d9b.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes/kubernetes.io~empty-dir/ca-trust-extracted/pem/directory-hash/869fbf79.0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/containers/registry/f8d22bdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c10,c16 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/6e8bbfac not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/54dd7996 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator/a4f1bb05 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/207129da not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/c1df39e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/containers/cluster-samples-operator-watch/15b8f1cd not reset as customized by admin to system_u:object_r:container_file_t:s0:c9,c17 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3523263858/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..2025_02_23_05_27_49.3256605594/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes/kubernetes.io~configmap/images/images.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/77bd6913 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/2382c1b1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/kube-rbac-proxy/704ce128 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/70d16fe0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/bfb95535 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/containers/machine-api-operator/57a8e8e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c0,c15 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..2025_02_23_05_27_49.3413793711/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/1b9d3e5e not reset as customized by admin to system_u:object_r:container_file_t:s0:c107,c917 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/fddb173c not reset as customized by admin to system_u:object_r:container_file_t:s0:c202,c983 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/containers/kube-apiserver-operator/95d3c6c4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c219,c404 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/bfb5fff5 not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/2aef40aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/9d751cbb-f2e2-430d-9754-c882a5e924a5/containers/check-endpoints/c0391cad not reset as customized by admin to system_u:object_r:container_file_t:s0:c20,c21 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/1119e69d not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/660608b4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager/8220bd53 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/85f99d5c not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/cluster-policy-controller/4b0225f6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/9c2a3394 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-cert-syncer/e820b243 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/1ca52ea0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c776,c1007 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/f614b9022728cf315e60c057852e563e/containers/kube-controller-manager-recovery-controller/e6988e45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c214,c928 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes/kubernetes.io~configmap/mcc-auth-proxy-config/..2025_02_24_06_09_21.2517297950/config-file.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/6655f00b not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/98bc3986 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/machine-config-controller/08e3458a not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/2a191cb0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/6c4eeefb not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/containers/kube-rbac-proxy/f61a549c not reset as customized by admin to system_u:object_r:container_file_t:s0:c4,c17 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/24891863 not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/hostpath-provisioner/fbdfd89c not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/9b63b3bc not reset as customized by admin to system_u:object_r:container_file_t:s0:c37,c572 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/liveness-probe/8acde6d6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/node-driver-registrar/59ecbba3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/containers/csi-provisioner/685d4be3 not reset as customized by admin to system_u:object_r:container_file_t:s0:c318,c553 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..2025_02_24_06_20_07.341639300/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/config.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.client-ca.configmap not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/config/openshift-route-controller-manager.serving-cert.secret not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851 not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..2025_02_24_06_20_07.2950937851/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes/kubernetes.io~configmap/client-ca/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/containers/route-controller-manager/feaea55e not reset as customized by admin to system_u:object_r:container_file_t:s0:c2,c23 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abinitio-runtime-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/accuknox-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aci-containers-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airlock-microgateway/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ako-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloy/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anchore-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-cloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/appdynamics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-dcap-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ccm-node-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cfm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cilium-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloud-native-postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudera-streams-messaging-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudnative-pg/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cnfv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/conjur-follower-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/coroot-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cte-k8s-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-deploy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/digitalai-release-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edb-hcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/elasticsearch-eck-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/federatorai-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fujitsu-enterprise-postgres-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/function-mesh/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/harness-gitops-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hcp-terraform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hpe-ezmeral-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-application-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-directory-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-dr-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-licensing-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infoscale-sds-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infrastructure-asset-orchestrator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-device-plugins-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/intel-kubernetes-power-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-openshift-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8s-triliovault/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-ati-updates/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-framework/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-ingress/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-licensing/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-kcos-sso/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-load-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-loadcore-agents/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nats-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-nimbusmosaic-dusim/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-rest-api-browser-v1/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-appsec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-db/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-diagnostics/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-logging/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-migration/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-msg-broker/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-notifications/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-stats-dashboards/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-storage/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-test-core/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-wap-ui/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keysight-websocket-service/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kong-gateway-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubearmor-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lenovo-locd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memcached-operator-ogaye/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/memory-machine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-enterprise/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netapp-spark-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-adm-agent-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netscaler-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-repository-ha-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nginx-ingress-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nim-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxiq-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nxrm-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odigos-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/open-liberty-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftartifactoryha-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshiftxray-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/operator-certification-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pmem-csi-operator-os/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-component-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/runtime-fabric-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sanstoragecsi-operator-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/smilecdr-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sriov-fec/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-commons-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stackable-zookeeper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-tsc-client-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tawon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tigera-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vcp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/webotx-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/63709497 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/d966b7fd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-utilities/f5773757 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/81c9edb9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/57bf57ee not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/extract-content/86f5e6aa not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/0aabe31d not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/d2af85c2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/containers/registry-server/09d157d9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/3scale-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-acmpca-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigateway-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-apigatewayv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-applicationautoscaling-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-athena-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudfront-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudtrail-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatch-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-cloudwatchlogs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-documentdb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-dynamodb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ec2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecr-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ecs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-efs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eks-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elasticache-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-elbv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-emrcontainers-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-eventbridge-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-iam-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kafka-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-keyspaces-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kinesis-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-kms-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-lambda-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-memorydb-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-mq-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-networkfirewall-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-opensearchservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-organizations-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-pipes-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-prometheusservice-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-rds-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-recyclebin-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-route53resolver-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-s3-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sagemaker-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-secretsmanager-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ses-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sfn-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sns-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-sqs-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-ssm-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ack-wafv2-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/airflow-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alloydb-omni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/alvearie-imaging-ingestion/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/amd-gpu-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/analytics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/annotationlab/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicast-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-api-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurio-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apicurito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/apimatic-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/application-services-metering-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aqua/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/argocd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/assisted-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/authorino-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/automotive-infra/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aws-efs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/awss3-operator-registry/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/azure-service-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/beegfs-csi-driver-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/bpfman-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-k/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/camel-karavan-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cass-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cert-utils-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-aas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-impairment-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cluster-manager/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/codeflare-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-kubevirt-hyperconverged/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-trivy-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/community-windows-machine-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/customized-user-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cxl-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dapr-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datatrucker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dbaas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/debezium-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dell-csm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/deployment-validation-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/devopsinabox/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-amlen-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eclipse-che/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ecr-secret-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/edp-keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eginnovations-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/egressip-ipam-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ember-csi-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/etcd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/eventing-kogito/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/external-secrets-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/falcon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fence-agents-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flink-kubernetes-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k8gb/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/fossul-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/github-arc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitops-primer/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/gitwebhook-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/global-load-balancer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/grafana-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/group-sync-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hawtio-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hazelcast-platform-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hedvig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hive-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/horreum-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/hyperfoil-bundle/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-block-csi-operator-community/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-security-verify-access-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibm-spectrum-scale-csi-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ibmcloud-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/infinispan/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/integrity-shield-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ipfs-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/istio-workspace-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/jaeger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kaoto-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keda/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keepalived-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/keycloak-permissions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/klusterlet/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kogito-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/koku-metrics-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/konveyor-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/korrel8r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kuadrant-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kube-green/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubecost/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubernetes-imagepuller-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/l5-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/layer7-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lbconfig-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/lib-bucket-provisioner/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/limitador-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/logging-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/loki-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/machine-deletion-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mariadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marin3r/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mercury-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/microcks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-atlas-kubernetes/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/mongodb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/move2kube-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multi-nic-cni-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-global-hub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/multicluster-operators-subscription/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/must-gather-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/namespace-configuration-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ncn-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ndmspc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/netobserv-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-community-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nexus-operator-m88i/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nfs-provisioner-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nlp-server/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-discovery-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-healthcheck-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/node-maintenance-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/nsm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oadp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/observability-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/oci-ccm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ocm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/odoo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opendatahub-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openebs/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-nfd-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-node-upgrade-mutex-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/openshift-qiskit-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/opentelemetry-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patch-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/patterns-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pcc-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pelorus-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/percona-xtradb-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/portworx-essentials/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/postgresql/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/proactive-node-scaling-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/project-quay/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometheus-exporter-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/prometurbo/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pubsubplus-eventbroker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pulp-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-cluster-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rabbitmq-messaging-topology-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/reportportal-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/resource-locker-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/rhoas-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ripsaw/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sailoperator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-commerce-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-data-intelligence-observer-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sap-hana-express-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/self-node-remediation/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/service-binding-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/shipwright-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sigstore-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/silicom-sts-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/skupper-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snapscheduler/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/snyk-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/socmmd/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonar-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosivio/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sonataflow-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/sosreport-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/spark-helm-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/special-resource-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/stolostron-engine/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/strimzi-kafka-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/syndesis/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tagger/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tempo-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tf-controller/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/tidb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trident-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/trustify-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ucs-ci-solutions-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/universal-crossplane/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/varnish-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vault-config-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/verticadb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/volume-expander-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/wandb-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/windup-operator/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yaks/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c0fe7256 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/c30319e4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-utilities/e6b1dd45 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/2bb643f0 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/920de426 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/extract-content/70fa1e87 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/a1c12a2f not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/9442e6c7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/containers/registry-server/5b45ec72 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/abot-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aerospike-kubernetes-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/aikit-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzo-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzograph-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/anzounstructured-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cloudbees-ci-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/cockroachdb-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/crunchy-postgres-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/datadog-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/dynatrace-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/entando-k8s-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/flux/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/instana-agent-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/iomesh-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/joget-dx8-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/k10-kasten-operator-term-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubemq-operator-marketplace-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/kubeturbo-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/linstor-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/marketplace-games-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/model-builder-for-vision-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/neuvector-certified-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/ovms-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/pachyderm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/redis-enterprise-operator-cert-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/seldon-deploy-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-paygo-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/starburst-enterprise-helm-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/t8c-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/timemachine-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/vfunction-server-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/xcrypt-operator-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/yugabyte-platform-operator-bundle-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/catalog/zabbix-operator-certified-rhmp/catalog.json not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/00000-1.psg.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/db.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/index.pmt not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/main.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/db/overflow.pix not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/catalog-content/cache/pogreb.v1/digest not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes/kubernetes.io~empty-dir/utilities/copy-content not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/3c9f3a59 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/1091c11b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-utilities/9a6821c6 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/ec0c35e2 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/517f37e7 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/extract-content/6214fe78 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/ba189c8b not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/351e4f31 not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/containers/registry-server/c0f219ff not reset as customized by admin to system_u:object_r:container_file_t:s0:c7,c13 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/8069f607 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/559c3d82 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/wait-for-host-port/605ad488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/148df488 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/3bf6dcb4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler/022a2feb not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/938c3924 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/729fe23e not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-cert-syncer/1fd5cbd4 not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/a96697e1 not reset as customized by admin to system_u:object_r:container_file_t:s0:c378,c723 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/e155ddca not reset as customized by admin to system_u:object_r:container_file_t:s0:c133,c223 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/3dcd261975c3d6b9a6ad6367fd4facd3/containers/kube-scheduler-recovery-controller/10dd0e0f not reset as customized by admin to system_u:object_r:container_file_t:s0:c247,c522 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..2025_02_24_06_09_35.3018472960/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-trusted-ca-bundle/ca-bundle.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..2025_02_24_06_09_35.4262376737/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/audit-policies/audit.yaml not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..2025_02_24_06_09_35.2630275752/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-cliconfig/v4-0-config-system-cliconfig not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..2025_02_24_06_09_35.2376963788/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/..data not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes/kubernetes.io~configmap/v4-0-config-system-service-ca/service-ca.crt not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/etc-hosts not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/6f2c8392 not reset as customized by admin to system_u:object_r:container_file_t:s0:c267,c588 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/containers/oauth-openshift/bd241ad9 not reset as customized by admin to system_u:object_r:container_file_t:s0:c682,c947 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/plugins not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/plugins/csi-hostpath not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/plugins/csi-hostpath/csi.sock not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/plugins/kubernetes.io not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/plugins/kubernetes.io/csi not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983 not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/vol_data.json not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 12 00:06:40 crc restorecon[4686]: /var/lib/kubelet/plugins_registry not reset as customized by admin to system_u:object_r:container_file_t:s0 Mar 12 00:06:40 crc restorecon[4686]: Relabeled /var/usrlocal/bin/kubenswrapper from system_u:object_r:bin_t:s0 to system_u:object_r:kubelet_exec_t:s0 Mar 12 00:06:41 crc kubenswrapper[4948]: Flag --container-runtime-endpoint has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 12 00:06:41 crc kubenswrapper[4948]: Flag --minimum-container-ttl-duration has been deprecated, Use --eviction-hard or --eviction-soft instead. Will be removed in a future version. Mar 12 00:06:41 crc kubenswrapper[4948]: Flag --volume-plugin-dir has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 12 00:06:41 crc kubenswrapper[4948]: Flag --register-with-taints has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 12 00:06:41 crc kubenswrapper[4948]: Flag --pod-infra-container-image has been deprecated, will be removed in a future release. Image garbage collector will get sandbox image information from CRI. Mar 12 00:06:41 crc kubenswrapper[4948]: Flag --system-reserved has been deprecated, This parameter should be set via the config file specified by the Kubelet's --config flag. See https://kubernetes.io/docs/tasks/administer-cluster/kubelet-config-file/ for more information. Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.035519 4948 server.go:211] "--pod-infra-container-image will not be pruned by the image garbage collector in kubelet and should also be set in the remote runtime" Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042070 4948 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042113 4948 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042129 4948 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042145 4948 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042161 4948 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042175 4948 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042189 4948 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042200 4948 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042212 4948 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042223 4948 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042233 4948 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042243 4948 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042254 4948 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042265 4948 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042288 4948 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042331 4948 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042343 4948 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042356 4948 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042368 4948 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042378 4948 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042409 4948 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042421 4948 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042431 4948 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042441 4948 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042452 4948 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042463 4948 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042474 4948 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042485 4948 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042495 4948 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042504 4948 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042513 4948 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042521 4948 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042529 4948 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042537 4948 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042545 4948 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042552 4948 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042560 4948 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042569 4948 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042576 4948 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042585 4948 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042593 4948 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042600 4948 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042609 4948 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042618 4948 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042626 4948 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042634 4948 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042643 4948 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042650 4948 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042659 4948 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042683 4948 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042691 4948 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042704 4948 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042713 4948 feature_gate.go:330] unrecognized feature gate: Example Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042725 4948 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042735 4948 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042745 4948 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042756 4948 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042766 4948 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042775 4948 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042784 4948 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042792 4948 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042799 4948 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042807 4948 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042815 4948 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042828 4948 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042839 4948 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042847 4948 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042856 4948 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042865 4948 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042874 4948 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.042882 4948 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.043810 4948 flags.go:64] FLAG: --address="0.0.0.0" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.043840 4948 flags.go:64] FLAG: --allowed-unsafe-sysctls="[]" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.043858 4948 flags.go:64] FLAG: --anonymous-auth="true" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.043871 4948 flags.go:64] FLAG: --application-metrics-count-limit="100" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.043884 4948 flags.go:64] FLAG: --authentication-token-webhook="false" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.043893 4948 flags.go:64] FLAG: --authentication-token-webhook-cache-ttl="2m0s" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.043905 4948 flags.go:64] FLAG: --authorization-mode="AlwaysAllow" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.043916 4948 flags.go:64] FLAG: --authorization-webhook-cache-authorized-ttl="5m0s" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.043925 4948 flags.go:64] FLAG: --authorization-webhook-cache-unauthorized-ttl="30s" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.043935 4948 flags.go:64] FLAG: --boot-id-file="/proc/sys/kernel/random/boot_id" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.043946 4948 flags.go:64] FLAG: --bootstrap-kubeconfig="/etc/kubernetes/kubeconfig" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.043956 4948 flags.go:64] FLAG: --cert-dir="/var/lib/kubelet/pki" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.043966 4948 flags.go:64] FLAG: --cgroup-driver="cgroupfs" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.043975 4948 flags.go:64] FLAG: --cgroup-root="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.043984 4948 flags.go:64] FLAG: --cgroups-per-qos="true" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.043993 4948 flags.go:64] FLAG: --client-ca-file="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044004 4948 flags.go:64] FLAG: --cloud-config="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044012 4948 flags.go:64] FLAG: --cloud-provider="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044021 4948 flags.go:64] FLAG: --cluster-dns="[]" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044032 4948 flags.go:64] FLAG: --cluster-domain="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044041 4948 flags.go:64] FLAG: --config="/etc/kubernetes/kubelet.conf" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044050 4948 flags.go:64] FLAG: --config-dir="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044059 4948 flags.go:64] FLAG: --container-hints="/etc/cadvisor/container_hints.json" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044068 4948 flags.go:64] FLAG: --container-log-max-files="5" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044079 4948 flags.go:64] FLAG: --container-log-max-size="10Mi" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044088 4948 flags.go:64] FLAG: --container-runtime-endpoint="/var/run/crio/crio.sock" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044098 4948 flags.go:64] FLAG: --containerd="/run/containerd/containerd.sock" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044107 4948 flags.go:64] FLAG: --containerd-namespace="k8s.io" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044116 4948 flags.go:64] FLAG: --contention-profiling="false" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044125 4948 flags.go:64] FLAG: --cpu-cfs-quota="true" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044134 4948 flags.go:64] FLAG: --cpu-cfs-quota-period="100ms" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044144 4948 flags.go:64] FLAG: --cpu-manager-policy="none" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044153 4948 flags.go:64] FLAG: --cpu-manager-policy-options="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044178 4948 flags.go:64] FLAG: --cpu-manager-reconcile-period="10s" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044187 4948 flags.go:64] FLAG: --enable-controller-attach-detach="true" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044196 4948 flags.go:64] FLAG: --enable-debugging-handlers="true" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044205 4948 flags.go:64] FLAG: --enable-load-reader="false" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044213 4948 flags.go:64] FLAG: --enable-server="true" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044222 4948 flags.go:64] FLAG: --enforce-node-allocatable="[pods]" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044235 4948 flags.go:64] FLAG: --event-burst="100" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044244 4948 flags.go:64] FLAG: --event-qps="50" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044253 4948 flags.go:64] FLAG: --event-storage-age-limit="default=0" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044263 4948 flags.go:64] FLAG: --event-storage-event-limit="default=0" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044272 4948 flags.go:64] FLAG: --eviction-hard="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044282 4948 flags.go:64] FLAG: --eviction-max-pod-grace-period="0" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044342 4948 flags.go:64] FLAG: --eviction-minimum-reclaim="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044352 4948 flags.go:64] FLAG: --eviction-pressure-transition-period="5m0s" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044361 4948 flags.go:64] FLAG: --eviction-soft="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044370 4948 flags.go:64] FLAG: --eviction-soft-grace-period="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044379 4948 flags.go:64] FLAG: --exit-on-lock-contention="false" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044389 4948 flags.go:64] FLAG: --experimental-allocatable-ignore-eviction="false" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044398 4948 flags.go:64] FLAG: --experimental-mounter-path="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044407 4948 flags.go:64] FLAG: --fail-cgroupv1="false" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044415 4948 flags.go:64] FLAG: --fail-swap-on="true" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044424 4948 flags.go:64] FLAG: --feature-gates="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044435 4948 flags.go:64] FLAG: --file-check-frequency="20s" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044444 4948 flags.go:64] FLAG: --global-housekeeping-interval="1m0s" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044454 4948 flags.go:64] FLAG: --hairpin-mode="promiscuous-bridge" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044463 4948 flags.go:64] FLAG: --healthz-bind-address="127.0.0.1" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044472 4948 flags.go:64] FLAG: --healthz-port="10248" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044481 4948 flags.go:64] FLAG: --help="false" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044491 4948 flags.go:64] FLAG: --hostname-override="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044499 4948 flags.go:64] FLAG: --housekeeping-interval="10s" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044508 4948 flags.go:64] FLAG: --http-check-frequency="20s" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044517 4948 flags.go:64] FLAG: --image-credential-provider-bin-dir="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044526 4948 flags.go:64] FLAG: --image-credential-provider-config="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044535 4948 flags.go:64] FLAG: --image-gc-high-threshold="85" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044544 4948 flags.go:64] FLAG: --image-gc-low-threshold="80" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044553 4948 flags.go:64] FLAG: --image-service-endpoint="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044562 4948 flags.go:64] FLAG: --kernel-memcg-notification="false" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044571 4948 flags.go:64] FLAG: --kube-api-burst="100" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044580 4948 flags.go:64] FLAG: --kube-api-content-type="application/vnd.kubernetes.protobuf" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044590 4948 flags.go:64] FLAG: --kube-api-qps="50" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044598 4948 flags.go:64] FLAG: --kube-reserved="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044608 4948 flags.go:64] FLAG: --kube-reserved-cgroup="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044616 4948 flags.go:64] FLAG: --kubeconfig="/var/lib/kubelet/kubeconfig" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044626 4948 flags.go:64] FLAG: --kubelet-cgroups="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044634 4948 flags.go:64] FLAG: --local-storage-capacity-isolation="true" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044643 4948 flags.go:64] FLAG: --lock-file="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044651 4948 flags.go:64] FLAG: --log-cadvisor-usage="false" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044660 4948 flags.go:64] FLAG: --log-flush-frequency="5s" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044669 4948 flags.go:64] FLAG: --log-json-info-buffer-size="0" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044689 4948 flags.go:64] FLAG: --log-json-split-stream="false" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044699 4948 flags.go:64] FLAG: --log-text-info-buffer-size="0" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044708 4948 flags.go:64] FLAG: --log-text-split-stream="false" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044719 4948 flags.go:64] FLAG: --logging-format="text" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044730 4948 flags.go:64] FLAG: --machine-id-file="/etc/machine-id,/var/lib/dbus/machine-id" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044744 4948 flags.go:64] FLAG: --make-iptables-util-chains="true" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044756 4948 flags.go:64] FLAG: --manifest-url="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044768 4948 flags.go:64] FLAG: --manifest-url-header="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044781 4948 flags.go:64] FLAG: --max-housekeeping-interval="15s" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044791 4948 flags.go:64] FLAG: --max-open-files="1000000" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044802 4948 flags.go:64] FLAG: --max-pods="110" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044811 4948 flags.go:64] FLAG: --maximum-dead-containers="-1" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044820 4948 flags.go:64] FLAG: --maximum-dead-containers-per-container="1" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044829 4948 flags.go:64] FLAG: --memory-manager-policy="None" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044837 4948 flags.go:64] FLAG: --minimum-container-ttl-duration="6m0s" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044847 4948 flags.go:64] FLAG: --minimum-image-ttl-duration="2m0s" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044856 4948 flags.go:64] FLAG: --node-ip="192.168.126.11" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044865 4948 flags.go:64] FLAG: --node-labels="node-role.kubernetes.io/control-plane=,node-role.kubernetes.io/master=,node.openshift.io/os_id=rhcos" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044884 4948 flags.go:64] FLAG: --node-status-max-images="50" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044893 4948 flags.go:64] FLAG: --node-status-update-frequency="10s" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044902 4948 flags.go:64] FLAG: --oom-score-adj="-999" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044911 4948 flags.go:64] FLAG: --pod-cidr="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044921 4948 flags.go:64] FLAG: --pod-infra-container-image="quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:33549946e22a9ffa738fd94b1345f90921bc8f92fa6137784cb33c77ad806f9d" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044940 4948 flags.go:64] FLAG: --pod-manifest-path="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044949 4948 flags.go:64] FLAG: --pod-max-pids="-1" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044958 4948 flags.go:64] FLAG: --pods-per-core="0" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044967 4948 flags.go:64] FLAG: --port="10250" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044976 4948 flags.go:64] FLAG: --protect-kernel-defaults="false" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044985 4948 flags.go:64] FLAG: --provider-id="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.044993 4948 flags.go:64] FLAG: --qos-reserved="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.045002 4948 flags.go:64] FLAG: --read-only-port="10255" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.045012 4948 flags.go:64] FLAG: --register-node="true" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.045021 4948 flags.go:64] FLAG: --register-schedulable="true" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.045030 4948 flags.go:64] FLAG: --register-with-taints="node-role.kubernetes.io/master=:NoSchedule" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.045043 4948 flags.go:64] FLAG: --registry-burst="10" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.045052 4948 flags.go:64] FLAG: --registry-qps="5" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.045061 4948 flags.go:64] FLAG: --reserved-cpus="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.045070 4948 flags.go:64] FLAG: --reserved-memory="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.045080 4948 flags.go:64] FLAG: --resolv-conf="/etc/resolv.conf" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.045089 4948 flags.go:64] FLAG: --root-dir="/var/lib/kubelet" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.045098 4948 flags.go:64] FLAG: --rotate-certificates="false" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.045108 4948 flags.go:64] FLAG: --rotate-server-certificates="false" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.045117 4948 flags.go:64] FLAG: --runonce="false" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.045125 4948 flags.go:64] FLAG: --runtime-cgroups="/system.slice/crio.service" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.045134 4948 flags.go:64] FLAG: --runtime-request-timeout="2m0s" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.045143 4948 flags.go:64] FLAG: --seccomp-default="false" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.045152 4948 flags.go:64] FLAG: --serialize-image-pulls="true" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.045161 4948 flags.go:64] FLAG: --storage-driver-buffer-duration="1m0s" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.045170 4948 flags.go:64] FLAG: --storage-driver-db="cadvisor" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.045179 4948 flags.go:64] FLAG: --storage-driver-host="localhost:8086" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.045188 4948 flags.go:64] FLAG: --storage-driver-password="root" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.045197 4948 flags.go:64] FLAG: --storage-driver-secure="false" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.045206 4948 flags.go:64] FLAG: --storage-driver-table="stats" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.045215 4948 flags.go:64] FLAG: --storage-driver-user="root" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.045224 4948 flags.go:64] FLAG: --streaming-connection-idle-timeout="4h0m0s" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.045236 4948 flags.go:64] FLAG: --sync-frequency="1m0s" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.045245 4948 flags.go:64] FLAG: --system-cgroups="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.045254 4948 flags.go:64] FLAG: --system-reserved="cpu=200m,ephemeral-storage=350Mi,memory=350Mi" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.045268 4948 flags.go:64] FLAG: --system-reserved-cgroup="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.045277 4948 flags.go:64] FLAG: --tls-cert-file="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.045285 4948 flags.go:64] FLAG: --tls-cipher-suites="[]" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.045297 4948 flags.go:64] FLAG: --tls-min-version="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.045335 4948 flags.go:64] FLAG: --tls-private-key-file="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.045344 4948 flags.go:64] FLAG: --topology-manager-policy="none" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.045353 4948 flags.go:64] FLAG: --topology-manager-policy-options="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.045363 4948 flags.go:64] FLAG: --topology-manager-scope="container" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.045372 4948 flags.go:64] FLAG: --v="2" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.045384 4948 flags.go:64] FLAG: --version="false" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.045395 4948 flags.go:64] FLAG: --vmodule="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.045405 4948 flags.go:64] FLAG: --volume-plugin-dir="/etc/kubernetes/kubelet-plugins/volume/exec" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.045415 4948 flags.go:64] FLAG: --volume-stats-agg-period="1m0s" Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.045609 4948 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.045620 4948 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.045628 4948 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.045637 4948 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.045645 4948 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.045653 4948 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.045661 4948 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.045669 4948 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.045677 4948 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.045685 4948 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.045692 4948 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.045700 4948 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.045708 4948 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.045716 4948 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.045726 4948 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.045735 4948 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.045757 4948 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.045767 4948 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.045776 4948 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.045784 4948 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.045792 4948 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.045799 4948 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.045807 4948 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.045816 4948 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.045830 4948 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.045838 4948 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.045845 4948 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.045853 4948 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.045861 4948 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.045872 4948 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.045882 4948 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.045890 4948 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.045901 4948 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.045911 4948 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.045920 4948 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.045929 4948 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.045937 4948 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.045946 4948 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.045954 4948 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.045962 4948 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.045970 4948 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.045979 4948 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.045987 4948 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.045996 4948 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.046004 4948 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.046014 4948 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.046024 4948 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.046032 4948 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.046043 4948 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.046051 4948 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.046061 4948 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.046070 4948 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.046079 4948 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.046089 4948 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.046097 4948 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.046105 4948 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.046116 4948 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.046124 4948 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.046132 4948 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.046142 4948 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.046151 4948 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.046159 4948 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.046168 4948 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.046175 4948 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.046183 4948 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.046193 4948 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.046200 4948 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.046208 4948 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.046215 4948 feature_gate.go:330] unrecognized feature gate: Example Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.046223 4948 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.046231 4948 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.047240 4948 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.060066 4948 server.go:491] "Kubelet version" kubeletVersion="v1.31.5" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.060125 4948 server.go:493] "Golang settings" GOGC="" GOMAXPROCS="" GOTRACEBACK="" Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.060282 4948 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.060329 4948 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.060341 4948 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.060355 4948 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.060367 4948 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.060377 4948 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.060387 4948 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.060397 4948 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.060407 4948 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.060417 4948 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.060427 4948 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.060437 4948 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.060447 4948 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.060457 4948 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.060470 4948 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.060481 4948 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.060491 4948 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.060501 4948 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.060511 4948 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.060521 4948 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.060531 4948 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.060540 4948 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.060550 4948 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.060560 4948 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.060573 4948 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.060586 4948 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.060596 4948 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.060606 4948 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.060616 4948 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.060627 4948 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.060637 4948 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.060646 4948 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.060656 4948 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.060665 4948 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.060676 4948 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.060687 4948 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.060698 4948 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.060709 4948 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.060719 4948 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.060729 4948 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.060739 4948 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.060749 4948 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.060758 4948 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.060768 4948 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.060782 4948 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.060797 4948 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.060809 4948 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.060821 4948 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.060831 4948 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.060841 4948 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.060854 4948 feature_gate.go:330] unrecognized feature gate: Example Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.060864 4948 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.060874 4948 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.060884 4948 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.060897 4948 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.060911 4948 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.060921 4948 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.060931 4948 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.060941 4948 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.060951 4948 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.060961 4948 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.060970 4948 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.060980 4948 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.060993 4948 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.061006 4948 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.061018 4948 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.061029 4948 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.061039 4948 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.061052 4948 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.061065 4948 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.061078 4948 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.061094 4948 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.061402 4948 feature_gate.go:330] unrecognized feature gate: RouteAdvertisements Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.061420 4948 feature_gate.go:330] unrecognized feature gate: BuildCSIVolumes Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.061431 4948 feature_gate.go:330] unrecognized feature gate: SetEIPForNLBIngressController Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.061442 4948 feature_gate.go:330] unrecognized feature gate: Example Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.061453 4948 feature_gate.go:330] unrecognized feature gate: NetworkSegmentation Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.061464 4948 feature_gate.go:330] unrecognized feature gate: GatewayAPI Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.061477 4948 feature_gate.go:330] unrecognized feature gate: ConsolePluginContentSecurityPolicy Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.061488 4948 feature_gate.go:330] unrecognized feature gate: ManagedBootImagesAWS Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.061499 4948 feature_gate.go:330] unrecognized feature gate: VolumeGroupSnapshot Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.061511 4948 feature_gate.go:330] unrecognized feature gate: MixedCPUsAllocation Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.061521 4948 feature_gate.go:330] unrecognized feature gate: ExternalOIDC Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.061531 4948 feature_gate.go:330] unrecognized feature gate: IngressControllerDynamicConfigurationManager Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.061541 4948 feature_gate.go:330] unrecognized feature gate: ImageStreamImportMode Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.061553 4948 feature_gate.go:351] Setting deprecated feature gate KMSv1=true. It will be removed in a future release. Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.061567 4948 feature_gate.go:330] unrecognized feature gate: SigstoreImageVerification Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.061577 4948 feature_gate.go:330] unrecognized feature gate: AWSEFSDriverVolumeMetrics Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.061587 4948 feature_gate.go:330] unrecognized feature gate: HardwareSpeed Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.061597 4948 feature_gate.go:330] unrecognized feature gate: OnClusterBuild Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.061607 4948 feature_gate.go:330] unrecognized feature gate: PrivateHostedZoneAWS Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.061617 4948 feature_gate.go:330] unrecognized feature gate: IngressControllerLBSubnetsAWS Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.061627 4948 feature_gate.go:330] unrecognized feature gate: VSphereMultiVCenters Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.061636 4948 feature_gate.go:330] unrecognized feature gate: AdditionalRoutingCapabilities Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.061646 4948 feature_gate.go:330] unrecognized feature gate: AWSClusterHostedDNS Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.061656 4948 feature_gate.go:330] unrecognized feature gate: PinnedImages Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.061666 4948 feature_gate.go:330] unrecognized feature gate: PlatformOperators Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.061675 4948 feature_gate.go:330] unrecognized feature gate: MachineAPIMigration Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.061682 4948 feature_gate.go:330] unrecognized feature gate: PersistentIPsForVirtualization Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.061692 4948 feature_gate.go:330] unrecognized feature gate: BootcNodeManagement Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.061702 4948 feature_gate.go:330] unrecognized feature gate: MinimumKubeletVersion Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.061713 4948 feature_gate.go:330] unrecognized feature gate: InsightsOnDemandDataGather Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.061723 4948 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAzure Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.061733 4948 feature_gate.go:330] unrecognized feature gate: NodeDisruptionPolicy Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.061743 4948 feature_gate.go:330] unrecognized feature gate: AutomatedEtcdBackup Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.061752 4948 feature_gate.go:330] unrecognized feature gate: ClusterMonitoringConfig Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.061764 4948 feature_gate.go:330] unrecognized feature gate: MultiArchInstallAWS Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.061775 4948 feature_gate.go:330] unrecognized feature gate: AlibabaPlatform Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.061786 4948 feature_gate.go:330] unrecognized feature gate: VSphereDriverConfiguration Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.061796 4948 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstallIBMCloud Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.061806 4948 feature_gate.go:330] unrecognized feature gate: NetworkLiveMigration Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.061815 4948 feature_gate.go:330] unrecognized feature gate: MachineConfigNodes Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.061826 4948 feature_gate.go:330] unrecognized feature gate: EtcdBackendQuota Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.061836 4948 feature_gate.go:330] unrecognized feature gate: AzureWorkloadIdentity Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.061845 4948 feature_gate.go:330] unrecognized feature gate: MachineAPIProviderOpenStack Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.061855 4948 feature_gate.go:330] unrecognized feature gate: ClusterAPIInstall Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.061865 4948 feature_gate.go:330] unrecognized feature gate: AdminNetworkPolicy Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.061876 4948 feature_gate.go:330] unrecognized feature gate: UpgradeStatus Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.061885 4948 feature_gate.go:330] unrecognized feature gate: MultiArchInstallGCP Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.061895 4948 feature_gate.go:330] unrecognized feature gate: GCPLabelsTags Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.061904 4948 feature_gate.go:330] unrecognized feature gate: BareMetalLoadBalancer Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.061939 4948 feature_gate.go:353] Setting GA feature gate ValidatingAdmissionPolicy=true. It will be removed in a future release. Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.061954 4948 feature_gate.go:330] unrecognized feature gate: VSphereControlPlaneMachineSet Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.061965 4948 feature_gate.go:330] unrecognized feature gate: OVNObservability Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.061977 4948 feature_gate.go:330] unrecognized feature gate: GCPClusterHostedDNS Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.061986 4948 feature_gate.go:330] unrecognized feature gate: OpenShiftPodSecurityAdmission Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.061996 4948 feature_gate.go:330] unrecognized feature gate: CSIDriverSharedResource Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.062009 4948 feature_gate.go:353] Setting GA feature gate DisableKubeletCloudCredentialProviders=true. It will be removed in a future release. Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.062021 4948 feature_gate.go:330] unrecognized feature gate: InsightsConfig Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.062033 4948 feature_gate.go:330] unrecognized feature gate: DNSNameResolver Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.062046 4948 feature_gate.go:353] Setting GA feature gate CloudDualStackNodeIPs=true. It will be removed in a future release. Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.062059 4948 feature_gate.go:330] unrecognized feature gate: SignatureStores Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.062071 4948 feature_gate.go:330] unrecognized feature gate: InsightsConfigAPI Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.062080 4948 feature_gate.go:330] unrecognized feature gate: VSphereMultiNetworks Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.062089 4948 feature_gate.go:330] unrecognized feature gate: MetricsCollectionProfiles Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.062097 4948 feature_gate.go:330] unrecognized feature gate: ManagedBootImages Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.062105 4948 feature_gate.go:330] unrecognized feature gate: VSphereStaticIPs Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.062113 4948 feature_gate.go:330] unrecognized feature gate: MachineAPIOperatorDisableMachineHealthCheckController Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.062121 4948 feature_gate.go:330] unrecognized feature gate: NewOLM Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.062129 4948 feature_gate.go:330] unrecognized feature gate: InsightsRuntimeExtractor Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.062137 4948 feature_gate.go:330] unrecognized feature gate: NutanixMultiSubnets Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.062145 4948 feature_gate.go:330] unrecognized feature gate: NetworkDiagnosticsConfig Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.062153 4948 feature_gate.go:330] unrecognized feature gate: ChunkSizeMiB Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.062167 4948 feature_gate.go:386] feature gates: {map[CloudDualStackNodeIPs:true DisableKubeletCloudCredentialProviders:true DynamicResourceAllocation:false EventedPLEG:false KMSv1:true MaxUnavailableStatefulSet:false NodeSwap:false ProcMountType:false RouteExternalCertificate:false ServiceAccountTokenNodeBinding:false TranslateStreamCloseWebsocketRequests:false UserNamespacesPodSecurityStandards:false UserNamespacesSupport:false ValidatingAdmissionPolicy:true VolumeAttributesClass:false]} Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.063200 4948 server.go:940] "Client rotation is on, will bootstrap in background" Mar 12 00:06:41 crc kubenswrapper[4948]: E0312 00:06:41.068739 4948 bootstrap.go:266] "Unhandled Error" err="part of the existing bootstrap client certificate in /var/lib/kubelet/kubeconfig is expired: 2026-02-24 05:52:08 +0000 UTC" logger="UnhandledError" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.074172 4948 bootstrap.go:101] "Use the bootstrap credentials to request a cert, and set kubeconfig to point to the certificate dir" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.074457 4948 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-client-current.pem". Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.078270 4948 server.go:997] "Starting client certificate rotation" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.078337 4948 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate rotation is enabled Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.078555 4948 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.108609 4948 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 12 00:06:41 crc kubenswrapper[4948]: E0312 00:06:41.113988 4948 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.180:6443: connect: connection refused" logger="UnhandledError" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.116169 4948 dynamic_cafile_content.go:161] "Starting controller" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.134090 4948 log.go:25] "Validated CRI v1 runtime API" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.177782 4948 log.go:25] "Validated CRI v1 image API" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.180500 4948 server.go:1437] "Using cgroup driver setting received from the CRI runtime" cgroupDriver="systemd" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.186757 4948 fs.go:133] Filesystem UUIDs: map[0b076daa-c26a-46d2-b3a6-72a8dbc6e257:/dev/vda4 2026-03-12-00-01-44-00:/dev/sr0 7B77-95E7:/dev/vda2 de0497b0-db1b-465a-b278-03db02455c71:/dev/vda3] Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.186814 4948 fs.go:134] Filesystem partitions: map[/dev/shm:{mountpoint:/dev/shm major:0 minor:22 fsType:tmpfs blockSize:0} /dev/vda3:{mountpoint:/boot major:252 minor:3 fsType:ext4 blockSize:0} /dev/vda4:{mountpoint:/var major:252 minor:4 fsType:xfs blockSize:0} /run:{mountpoint:/run major:0 minor:24 fsType:tmpfs blockSize:0} /run/user/1000:{mountpoint:/run/user/1000 major:0 minor:42 fsType:tmpfs blockSize:0} /tmp:{mountpoint:/tmp major:0 minor:30 fsType:tmpfs blockSize:0} /var/lib/etcd:{mountpoint:/var/lib/etcd major:0 minor:43 fsType:tmpfs blockSize:0}] Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.212737 4948 manager.go:217] Machine: {Timestamp:2026-03-12 00:06:41.210235212 +0000 UTC m=+0.665839030 CPUVendorID:AuthenticAMD NumCores:12 NumPhysicalCores:1 NumSockets:12 CpuFrequency:2799998 MemoryCapacity:33654132736 SwapCapacity:0 MemoryByType:map[] NVMInfo:{MemoryModeCapacity:0 AppDirectModeCapacity:0 AvgPowerBudget:0} HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] MachineID:21801e6708c44f15b81395eb736a7cec SystemUUID:b18e4bf2-d37e-4b80-8da3-cd3803754ba2 BootID:17825588-29cf-4059-bab7-1b042b9bf1f3 Filesystems:[{Device:/dev/shm DeviceMajor:0 DeviceMinor:22 Capacity:16827064320 Type:vfs Inodes:4108170 HasInodes:true} {Device:/run DeviceMajor:0 DeviceMinor:24 Capacity:6730829824 Type:vfs Inodes:819200 HasInodes:true} {Device:/dev/vda4 DeviceMajor:252 DeviceMinor:4 Capacity:85292941312 Type:vfs Inodes:41679680 HasInodes:true} {Device:/tmp DeviceMajor:0 DeviceMinor:30 Capacity:16827068416 Type:vfs Inodes:1048576 HasInodes:true} {Device:/dev/vda3 DeviceMajor:252 DeviceMinor:3 Capacity:366869504 Type:vfs Inodes:98304 HasInodes:true} {Device:/run/user/1000 DeviceMajor:0 DeviceMinor:42 Capacity:3365412864 Type:vfs Inodes:821634 HasInodes:true} {Device:/var/lib/etcd DeviceMajor:0 DeviceMinor:43 Capacity:1073741824 Type:vfs Inodes:4108170 HasInodes:true}] DiskMap:map[252:0:{Name:vda Major:252 Minor:0 Size:214748364800 Scheduler:none}] NetworkDevices:[{Name:br-ex MacAddress:fa:16:3e:80:82:75 Speed:0 Mtu:1500} {Name:br-int MacAddress:d6:39:55:2e:22:71 Speed:0 Mtu:1400} {Name:ens3 MacAddress:fa:16:3e:80:82:75 Speed:-1 Mtu:1500} {Name:ens7 MacAddress:fa:16:3e:14:3f:1c Speed:-1 Mtu:1500} {Name:ens7.20 MacAddress:52:54:00:10:0e:d1 Speed:-1 Mtu:1496} {Name:ens7.21 MacAddress:52:54:00:d0:66:68 Speed:-1 Mtu:1496} {Name:ens7.22 MacAddress:52:54:00:89:46:13 Speed:-1 Mtu:1496} {Name:eth10 MacAddress:b2:39:31:93:3e:4e Speed:0 Mtu:1500} {Name:ovn-k8s-mp0 MacAddress:0a:58:0a:d9:00:02 Speed:0 Mtu:1400} {Name:ovs-system MacAddress:12:1f:ce:cb:f7:d5 Speed:0 Mtu:1500}] Topology:[{Id:0 Memory:33654132736 HugePages:[{PageSize:1048576 NumPages:0} {PageSize:2048 NumPages:0}] Cores:[{Id:0 Threads:[0] Caches:[{Id:0 Size:32768 Type:Data Level:1} {Id:0 Size:32768 Type:Instruction Level:1} {Id:0 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:0 Size:16777216 Type:Unified Level:3}] SocketID:0 BookID: DrawerID:} {Id:0 Threads:[1] Caches:[{Id:1 Size:32768 Type:Data Level:1} {Id:1 Size:32768 Type:Instruction Level:1} {Id:1 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:1 Size:16777216 Type:Unified Level:3}] SocketID:1 BookID: DrawerID:} {Id:0 Threads:[10] Caches:[{Id:10 Size:32768 Type:Data Level:1} {Id:10 Size:32768 Type:Instruction Level:1} {Id:10 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:10 Size:16777216 Type:Unified Level:3}] SocketID:10 BookID: DrawerID:} {Id:0 Threads:[11] Caches:[{Id:11 Size:32768 Type:Data Level:1} {Id:11 Size:32768 Type:Instruction Level:1} {Id:11 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:11 Size:16777216 Type:Unified Level:3}] SocketID:11 BookID: DrawerID:} {Id:0 Threads:[2] Caches:[{Id:2 Size:32768 Type:Data Level:1} {Id:2 Size:32768 Type:Instruction Level:1} {Id:2 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:2 Size:16777216 Type:Unified Level:3}] SocketID:2 BookID: DrawerID:} {Id:0 Threads:[3] Caches:[{Id:3 Size:32768 Type:Data Level:1} {Id:3 Size:32768 Type:Instruction Level:1} {Id:3 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:3 Size:16777216 Type:Unified Level:3}] SocketID:3 BookID: DrawerID:} {Id:0 Threads:[4] Caches:[{Id:4 Size:32768 Type:Data Level:1} {Id:4 Size:32768 Type:Instruction Level:1} {Id:4 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:4 Size:16777216 Type:Unified Level:3}] SocketID:4 BookID: DrawerID:} {Id:0 Threads:[5] Caches:[{Id:5 Size:32768 Type:Data Level:1} {Id:5 Size:32768 Type:Instruction Level:1} {Id:5 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:5 Size:16777216 Type:Unified Level:3}] SocketID:5 BookID: DrawerID:} {Id:0 Threads:[6] Caches:[{Id:6 Size:32768 Type:Data Level:1} {Id:6 Size:32768 Type:Instruction Level:1} {Id:6 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:6 Size:16777216 Type:Unified Level:3}] SocketID:6 BookID: DrawerID:} {Id:0 Threads:[7] Caches:[{Id:7 Size:32768 Type:Data Level:1} {Id:7 Size:32768 Type:Instruction Level:1} {Id:7 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:7 Size:16777216 Type:Unified Level:3}] SocketID:7 BookID: DrawerID:} {Id:0 Threads:[8] Caches:[{Id:8 Size:32768 Type:Data Level:1} {Id:8 Size:32768 Type:Instruction Level:1} {Id:8 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:8 Size:16777216 Type:Unified Level:3}] SocketID:8 BookID: DrawerID:} {Id:0 Threads:[9] Caches:[{Id:9 Size:32768 Type:Data Level:1} {Id:9 Size:32768 Type:Instruction Level:1} {Id:9 Size:524288 Type:Unified Level:2}] UncoreCaches:[{Id:9 Size:16777216 Type:Unified Level:3}] SocketID:9 BookID: DrawerID:}] Caches:[] Distances:[10]}] CloudProvider:Unknown InstanceType:Unknown InstanceID:None} Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.213263 4948 manager_no_libpfm.go:29] cAdvisor is build without cgo and/or libpfm support. Perf event counters are not available. Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.213643 4948 manager.go:233] Version: {KernelVersion:5.14.0-427.50.2.el9_4.x86_64 ContainerOsVersion:Red Hat Enterprise Linux CoreOS 418.94.202502100215-0 DockerVersion: DockerAPIVersion: CadvisorVersion: CadvisorRevision:} Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.214490 4948 swap_util.go:113] "Swap is on" /proc/swaps contents="Filename\t\t\t\tType\t\tSize\t\tUsed\t\tPriority" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.214834 4948 container_manager_linux.go:267] "Container manager verified user specified cgroup-root exists" cgroupRoot=[] Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.214893 4948 container_manager_linux.go:272] "Creating Container Manager object based on Node Config" nodeConfig={"NodeName":"crc","RuntimeCgroupsName":"/system.slice/crio.service","SystemCgroupsName":"/system.slice","KubeletCgroupsName":"","KubeletOOMScoreAdj":-999,"ContainerRuntime":"","CgroupsPerQOS":true,"CgroupRoot":"/","CgroupDriver":"systemd","KubeletRootDir":"/var/lib/kubelet","ProtectKernelDefaults":true,"KubeReservedCgroupName":"","SystemReservedCgroupName":"","ReservedSystemCPUs":{},"EnforceNodeAllocatable":{"pods":{}},"KubeReserved":null,"SystemReserved":{"cpu":"200m","ephemeral-storage":"350Mi","memory":"350Mi"},"HardEvictionThresholds":[{"Signal":"nodefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.15},"GracePeriod":0,"MinReclaim":null},{"Signal":"imagefs.inodesFree","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.05},"GracePeriod":0,"MinReclaim":null},{"Signal":"memory.available","Operator":"LessThan","Value":{"Quantity":"100Mi","Percentage":0},"GracePeriod":0,"MinReclaim":null},{"Signal":"nodefs.available","Operator":"LessThan","Value":{"Quantity":null,"Percentage":0.1},"GracePeriod":0,"MinReclaim":null}],"QOSReserved":{},"CPUManagerPolicy":"none","CPUManagerPolicyOptions":null,"TopologyManagerScope":"container","CPUManagerReconcilePeriod":10000000000,"ExperimentalMemoryManagerPolicy":"None","ExperimentalMemoryManagerReservedMemory":null,"PodPidsLimit":4096,"EnforceCPULimits":true,"CPUCFSQuotaPeriod":100000000,"TopologyManagerPolicy":"none","TopologyManagerPolicyOptions":null,"CgroupVersion":2} Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.215252 4948 topology_manager.go:138] "Creating topology manager with none policy" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.215273 4948 container_manager_linux.go:303] "Creating device plugin manager" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.215954 4948 manager.go:142] "Creating Device Plugin manager" path="/var/lib/kubelet/device-plugins/kubelet.sock" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.216008 4948 server.go:66] "Creating device plugin registration server" version="v1beta1" socket="/var/lib/kubelet/device-plugins/kubelet.sock" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.216254 4948 state_mem.go:36] "Initialized new in-memory state store" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.216817 4948 server.go:1245] "Using root directory" path="/var/lib/kubelet" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.220760 4948 kubelet.go:418] "Attempting to sync node with API server" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.220799 4948 kubelet.go:313] "Adding static pod path" path="/etc/kubernetes/manifests" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.220841 4948 file.go:69] "Watching path" path="/etc/kubernetes/manifests" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.220863 4948 kubelet.go:324] "Adding apiserver pod source" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.220882 4948 apiserver.go:42] "Waiting for node sync before watching apiserver pods" Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.228552 4948 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.180:6443: connect: connection refused Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.228572 4948 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.180:6443: connect: connection refused Mar 12 00:06:41 crc kubenswrapper[4948]: E0312 00:06:41.229561 4948 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.180:6443: connect: connection refused" logger="UnhandledError" Mar 12 00:06:41 crc kubenswrapper[4948]: E0312 00:06:41.229565 4948 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.180:6443: connect: connection refused" logger="UnhandledError" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.230800 4948 kuberuntime_manager.go:262] "Container runtime initialized" containerRuntime="cri-o" version="1.31.5-4.rhaos4.18.gitdad78d5.el9" apiVersion="v1" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.232249 4948 certificate_store.go:130] Loading cert/key pair from "/var/lib/kubelet/pki/kubelet-server-current.pem". Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.236242 4948 kubelet.go:854] "Not starting ClusterTrustBundle informer because we are in static kubelet mode" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.237989 4948 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/portworx-volume" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.238037 4948 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/empty-dir" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.238054 4948 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/git-repo" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.238068 4948 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/host-path" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.238090 4948 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/nfs" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.238103 4948 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/secret" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.238118 4948 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/iscsi" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.238139 4948 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/downward-api" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.238156 4948 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/fc" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.238171 4948 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/configmap" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.238189 4948 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/projected" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.238202 4948 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/local-volume" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.239186 4948 plugins.go:603] "Loaded volume plugin" pluginName="kubernetes.io/csi" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.239851 4948 server.go:1280] "Started kubelet" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.240031 4948 server.go:163] "Starting to listen" address="0.0.0.0" port=10250 Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.241044 4948 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.180:6443: connect: connection refused Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.241366 4948 ratelimit.go:55] "Setting rate limiting for endpoint" service="podresources" qps=100 burstTokens=10 Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.242095 4948 server.go:236] "Starting to serve the podresources API" endpoint="unix:/var/lib/kubelet/pod-resources/kubelet.sock" Mar 12 00:06:41 crc systemd[1]: Started Kubernetes Kubelet. Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.243047 4948 server.go:460] "Adding debug handlers to kubelet server" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.244146 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate rotation is enabled Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.244191 4948 fs_resource_analyzer.go:67] "Starting FS ResourceAnalyzer" Mar 12 00:06:41 crc kubenswrapper[4948]: E0312 00:06:41.244515 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.244545 4948 volume_manager.go:287] "The desired_state_of_world populator starts" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.244566 4948 volume_manager.go:289] "Starting Kubelet Volume Manager" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.244675 4948 desired_state_of_world_populator.go:146] "Desired state populator starts to run" Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.245078 4948 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.180:6443: connect: connection refused Mar 12 00:06:41 crc kubenswrapper[4948]: E0312 00:06:41.245148 4948 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.180:6443: connect: connection refused" logger="UnhandledError" Mar 12 00:06:41 crc kubenswrapper[4948]: E0312 00:06:41.246648 4948 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.180:6443: connect: connection refused" interval="200ms" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.249270 4948 factory.go:219] Registration of the containerd container factory failed: unable to create containerd client: containerd: cannot unix dial containerd api service: dial unix /run/containerd/containerd.sock: connect: no such file or directory Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.249332 4948 factory.go:55] Registering systemd factory Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.249347 4948 factory.go:221] Registration of the systemd container factory successfully Mar 12 00:06:41 crc kubenswrapper[4948]: E0312 00:06:41.248175 4948 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": dial tcp 38.102.83.180:6443: connect: connection refused" event="&Event{ObjectMeta:{crc.189bef4f15559a92 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:41.239808658 +0000 UTC m=+0.695412426,LastTimestamp:2026-03-12 00:06:41.239808658 +0000 UTC m=+0.695412426,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.249777 4948 factory.go:153] Registering CRI-O factory Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.249828 4948 factory.go:221] Registration of the crio container factory successfully Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.249866 4948 factory.go:103] Registering Raw factory Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.249894 4948 manager.go:1196] Started watching for new ooms in manager Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.251486 4948 manager.go:319] Starting recovery of all containers Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.265388 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.265526 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.265551 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.265605 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.265625 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.265646 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.265674 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.265695 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.270059 4948 reconstruct.go:144] "Volume is marked device as uncertain and added into the actual state" volumeName="kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" deviceMountPath="/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.270135 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.270164 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.270185 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.270204 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.270224 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.270248 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="37a5e44f-9a88-4405-be8a-b645485e7312" volumeName="kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.270281 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.270335 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.270354 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.270380 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.270400 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.270426 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.270455 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.270474 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.270496 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.270514 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.270532 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.270553 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.270576 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.270596 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.270619 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.270639 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.270659 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.270685 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.270711 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.270752 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.270772 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="44663579-783b-4372-86d6-acf235a62d72" volumeName="kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.270831 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.270852 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.270872 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.270890 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.270914 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.270933 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3ab1a177-2de0-46d9-b765-d0d0649bb42e" volumeName="kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.270951 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.270968 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.271017 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.271038 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.271055 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="d75a4c96-2883-4a0b-bab2-0fab2b6c0b49" volumeName="kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.271074 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.271094 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.271112 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.271130 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.271147 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.271166 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.271193 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.271212 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.271233 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.271282 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" volumeName="kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.271328 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.271351 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.271396 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.271418 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.271437 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.271456 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.271474 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.271494 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.271513 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.271532 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.271553 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.271573 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.271590 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.271610 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.271629 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.271649 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.271667 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.271686 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.271709 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.271729 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.271747 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.271765 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.271784 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.271803 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.271857 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.271875 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.271900 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.271921 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.271940 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.271959 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="efdd0498-1daa-4136-9a4a-3b948c2293fc" volumeName="kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.271976 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.271995 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.272012 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7bb08738-c794-4ee8-9972-3a62ca171029" volumeName="kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.272033 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.272052 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.272070 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.272089 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.272107 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.272127 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.272146 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.272164 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d751cbb-f2e2-430d-9754-c882a5e924a5" volumeName="kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.272182 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.272201 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.272219 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.272236 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" volumeName="kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.272255 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.272272 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="496e6271-fb68-4057-954e-a0d97a4afa3f" volumeName="kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.272290 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.272340 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.272360 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.272381 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.272401 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" volumeName="kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.272425 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.272445 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.272467 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.272490 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.272510 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.272530 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.272549 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.272568 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.272586 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.272606 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.272627 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.272646 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.272665 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.272684 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6402fda4-df10-493c-b4e5-d0569419652d" volumeName="kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.272704 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.272726 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.272743 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.272761 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.272780 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.272797 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.272816 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.272835 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.272855 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" volumeName="kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.272872 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.272891 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" volumeName="kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.272910 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.272928 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5441d097-087c-4d9a-baa8-b210afa90fc9" volumeName="kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.272946 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6ea678ab-3438-413e-bfe3-290ae7725660" volumeName="kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.272964 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="c03ee662-fb2f-4fc4-a2c1-af487c19d254" volumeName="kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.272982 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.273003 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.273022 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.273040 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49ef4625-1d3a-4a9f-b595-c2433d32326d" volumeName="kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.273058 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6731426b-95fe-49ff-bb5f-40441049fde2" volumeName="kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.273076 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7539238d-5fe0-46ed-884e-1c3b566537ec" volumeName="kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.273094 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.273140 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.273159 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="43509403-f426-496e-be36-56cef71462f5" volumeName="kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.273178 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.273196 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.273213 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.273230 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.273250 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" volumeName="kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.273268 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" volumeName="kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.273348 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.273368 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.273389 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.273408 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.273427 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.273446 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.273465 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09efc573-dbb6-4249-bd59-9b87aba8dd28" volumeName="kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.273483 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.273504 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" volumeName="kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.273524 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" volumeName="kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.273542 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b78653f-4ff9-4508-8672-245ed9b561e3" volumeName="kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.273561 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.273580 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="87cf06ed-a83f-41a7-828d-70653580a8cb" volumeName="kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.273599 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.273627 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5225d0e4-402f-4861-b410-819f433b1803" volumeName="kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.273646 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.273666 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.273684 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" volumeName="kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.273704 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="9d4552c7-cd75-42dd-8880-30dd377c49a4" volumeName="kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.273722 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6312bbd-5731-4ea0-a20f-81d5a57df44a" volumeName="kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.273740 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.273758 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.273777 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" volumeName="kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.273794 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="4bb40260-dbaa-4fb0-84df-5e680505d512" volumeName="kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.273813 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="57a731c4-ef35-47a8-b875-bfb08a7f8011" volumeName="kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.273830 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="7583ce53-e0fe-4a16-9e4d-50516596a136" volumeName="kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.273847 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.273872 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="bf126b07-da06-4140-9a57-dfd54fc6b486" volumeName="kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.273892 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" volumeName="kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.273910 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1d611f23-29be-4491-8495-bee1670e935f" volumeName="kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.273927 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="31d8b7a1-420e-4252-a5b7-eebe8a111292" volumeName="kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.273945 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.273964 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="e7e6199b-1264-4501-8953-767f51328d08" volumeName="kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.273990 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="ef543e1b-8068-4ea3-b32a-61027b32e95d" volumeName="kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.274010 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.274028 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="0b574797-001e-440a-8f4e-c0be86edad0f" volumeName="kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.274046 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="25e176fe-21b4-4974-b1ed-c8b94f112a7f" volumeName="kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.274065 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="fda69060-fa79-4696-b1a6-7980f124bf7c" volumeName="kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.274084 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1bf7eb37-55a3-4c65-b768-a94c82151e69" volumeName="kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.274101 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.274121 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.274175 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5fe579f8-e8a6-4643-bce5-a661393c4dde" volumeName="kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.274192 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="6509e943-70c6-444c-bc41-48a544e36fbd" volumeName="kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.274209 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" volumeName="kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.274229 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="1386a44e-36a2-460c-96d0-0359d2b6f0f5" volumeName="kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.274247 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="3b6479f0-333b-4a96-9adf-2099afdc2447" volumeName="kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.274265 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="22c825df-677d-4ca6-82db-3454ed06e783" volumeName="kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.274284 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" volumeName="kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.274326 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="5b88f790-22fa-440e-b583-365168c0b23d" volumeName="kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.274344 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="8f668bae-612b-4b75-9490-919e737c6a3b" volumeName="kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.274363 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="925f1c65-6136-48ba-85aa-3a3b50560753" volumeName="kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.274381 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" volumeName="kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.274400 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="01ab3dd5-8196-46d0-ad33-122e2ca51def" volumeName="kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.274418 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="20b0d48f-5fd6-431c-a545-e3c800c7b866" volumeName="kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.274437 4948 reconstruct.go:130] "Volume is marked as uncertain and added into the actual state" pod="" podName="a31745f5-9847-4afe-82a5-3161cc66ca93" volumeName="kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" seLinuxMountContext="" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.274455 4948 reconstruct.go:97] "Volume reconstruction finished" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.274469 4948 reconciler.go:26] "Reconciler: start to sync state" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.289662 4948 manager.go:324] Recovery completed Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.306599 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.309139 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.309127 4948 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv4" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.309199 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.309436 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.311020 4948 cpu_manager.go:225] "Starting CPU manager" policy="none" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.311050 4948 cpu_manager.go:226] "Reconciling" reconcilePeriod="10s" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.311150 4948 state_mem.go:36] "Initialized new in-memory state store" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.311163 4948 kubelet_network_linux.go:50] "Initialized iptables rules." protocol="IPv6" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.311230 4948 status_manager.go:217] "Starting to sync pod status with apiserver" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.311263 4948 kubelet.go:2335] "Starting kubelet main sync loop" Mar 12 00:06:41 crc kubenswrapper[4948]: E0312 00:06:41.311379 4948 kubelet.go:2359] "Skipping pod synchronization" err="[container runtime status check may not have completed yet, PLEG is not healthy: pleg has yet to be successful]" Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.313111 4948 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.180:6443: connect: connection refused Mar 12 00:06:41 crc kubenswrapper[4948]: E0312 00:06:41.313192 4948 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.180:6443: connect: connection refused" logger="UnhandledError" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.328641 4948 policy_none.go:49] "None policy: Start" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.331105 4948 memory_manager.go:170] "Starting memorymanager" policy="None" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.331158 4948 state_mem.go:35] "Initializing new in-memory state store" Mar 12 00:06:41 crc kubenswrapper[4948]: E0312 00:06:41.344988 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.402068 4948 manager.go:334] "Starting Device Plugin manager" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.402152 4948 manager.go:513] "Failed to read data from checkpoint" checkpoint="kubelet_internal_checkpoint" err="checkpoint is not found" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.402173 4948 server.go:79] "Starting device plugin registration server" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.402845 4948 eviction_manager.go:189] "Eviction manager: starting control loop" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.402882 4948 container_log_manager.go:189] "Initializing container log rotate workers" workers=1 monitorPeriod="10s" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.403089 4948 plugin_watcher.go:51] "Plugin Watcher Start" path="/var/lib/kubelet/plugins_registry" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.403222 4948 plugin_manager.go:116] "The desired_state_of_world populator (plugin watcher) starts" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.403236 4948 plugin_manager.go:118] "Starting Kubelet Plugin Manager" Mar 12 00:06:41 crc kubenswrapper[4948]: E0312 00:06:41.409372 4948 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.411550 4948 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-controller-manager/kube-controller-manager-crc","openshift-kube-scheduler/openshift-kube-scheduler-crc","openshift-machine-config-operator/kube-rbac-proxy-crio-crc","openshift-etcd/etcd-crc","openshift-kube-apiserver/kube-apiserver-crc"] Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.411652 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.413036 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.413084 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.413107 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.413297 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.413770 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.413839 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.414363 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.414416 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.414435 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.414589 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.414738 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.414790 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.415101 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.415178 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.415349 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.417597 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.417646 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.417652 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.417663 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.417675 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.417688 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.417849 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.418165 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.418266 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.419163 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.419193 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.419205 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.419381 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.419579 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.419644 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.422480 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.422513 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.422515 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.422527 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.422556 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.422581 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.422709 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.422727 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.422786 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.422880 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.422927 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.426767 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.426958 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.427104 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:06:41 crc kubenswrapper[4948]: E0312 00:06:41.447647 4948 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.180:6443: connect: connection refused" interval="400ms" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.476629 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.476705 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.476738 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.476768 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.476866 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.476946 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.476998 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.477045 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.477180 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.477224 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.477259 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.477334 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.477395 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.477424 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.477486 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.503016 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.504262 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.504376 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.504398 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.504439 4948 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 12 00:06:41 crc kubenswrapper[4948]: E0312 00:06:41.505169 4948 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.180:6443: connect: connection refused" node="crc" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.578888 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.578949 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.578984 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.579014 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.579045 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.579103 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-cert-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.579134 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-cert-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.579135 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.579200 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-cert-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.579196 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-log-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.579254 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.579355 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.579272 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-var-lib-kubelet\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.579330 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f614b9022728cf315e60c057852e563e-resource-dir\") pod \"kube-controller-manager-crc\" (UID: \"f614b9022728cf315e60c057852e563e\") " pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.579456 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.579491 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.579525 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.579549 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"data-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-data-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.579588 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.579562 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"static-pod-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-static-pod-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.579771 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.579660 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.579836 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.579665 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kube\" (UniqueName: \"kubernetes.io/host-path/d1b160f5dda77d281dd8e69ec8d817f9-etc-kube\") pod \"kube-rbac-proxy-crio-crc\" (UID: \"d1b160f5dda77d281dd8e69ec8d817f9\") " pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.579877 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.579892 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"usr-local-bin\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-usr-local-bin\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.579869 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/2139d3e2895fc6797b9c76a1b4c9886d-resource-dir\") pod \"etcd-crc\" (UID: \"2139d3e2895fc6797b9c76a1b4c9886d\") " pod="openshift-etcd/etcd-crc" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.579927 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.579963 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.580016 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/3dcd261975c3d6b9a6ad6367fd4facd3-resource-dir\") pod \"openshift-kube-scheduler-crc\" (UID: \"3dcd261975c3d6b9a6ad6367fd4facd3\") " pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.706197 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.708074 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.708122 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.708134 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.708160 4948 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 12 00:06:41 crc kubenswrapper[4948]: E0312 00:06:41.708663 4948 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.180:6443: connect: connection refused" node="crc" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.778109 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.789136 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.812378 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.823774 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd/etcd-crc" Mar 12 00:06:41 crc kubenswrapper[4948]: I0312 00:06:41.831956 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.832597 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3dcd261975c3d6b9a6ad6367fd4facd3.slice/crio-c7852645f8648eaca077097c34cb29cc20a72599c7c0b4728c01211b97bb57c0 WatchSource:0}: Error finding container c7852645f8648eaca077097c34cb29cc20a72599c7c0b4728c01211b97bb57c0: Status 404 returned error can't find the container with id c7852645f8648eaca077097c34cb29cc20a72599c7c0b4728c01211b97bb57c0 Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.834095 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf614b9022728cf315e60c057852e563e.slice/crio-47f8ff230f33182efc550c78b233e370f2a82585c2c8d4d806fbc3d4082cffdc WatchSource:0}: Error finding container 47f8ff230f33182efc550c78b233e370f2a82585c2c8d4d806fbc3d4082cffdc: Status 404 returned error can't find the container with id 47f8ff230f33182efc550c78b233e370f2a82585c2c8d4d806fbc3d4082cffdc Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.844965 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd1b160f5dda77d281dd8e69ec8d817f9.slice/crio-b6d0e1098581c5d959aceffd999a879bdad99e876db181a5c2c00bb1b5dfa153 WatchSource:0}: Error finding container b6d0e1098581c5d959aceffd999a879bdad99e876db181a5c2c00bb1b5dfa153: Status 404 returned error can't find the container with id b6d0e1098581c5d959aceffd999a879bdad99e876db181a5c2c00bb1b5dfa153 Mar 12 00:06:41 crc kubenswrapper[4948]: E0312 00:06:41.848509 4948 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.180:6443: connect: connection refused" interval="800ms" Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.852217 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2139d3e2895fc6797b9c76a1b4c9886d.slice/crio-642dbb27b3c8456c103813e3ea308da2d30af88a3f6ef9a953399ae1a585c7cc WatchSource:0}: Error finding container 642dbb27b3c8456c103813e3ea308da2d30af88a3f6ef9a953399ae1a585c7cc: Status 404 returned error can't find the container with id 642dbb27b3c8456c103813e3ea308da2d30af88a3f6ef9a953399ae1a585c7cc Mar 12 00:06:41 crc kubenswrapper[4948]: W0312 00:06:41.861012 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf4b27818a5e8e43d0dc095d08835c792.slice/crio-91bc95ec8f59e0db090f0580cb985cb6fb9f01fde5d6de934aa0c096b3c6f280 WatchSource:0}: Error finding container 91bc95ec8f59e0db090f0580cb985cb6fb9f01fde5d6de934aa0c096b3c6f280: Status 404 returned error can't find the container with id 91bc95ec8f59e0db090f0580cb985cb6fb9f01fde5d6de934aa0c096b3c6f280 Mar 12 00:06:42 crc kubenswrapper[4948]: I0312 00:06:42.108794 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:06:42 crc kubenswrapper[4948]: I0312 00:06:42.110548 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:06:42 crc kubenswrapper[4948]: I0312 00:06:42.110588 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:06:42 crc kubenswrapper[4948]: I0312 00:06:42.110597 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:06:42 crc kubenswrapper[4948]: I0312 00:06:42.110619 4948 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 12 00:06:42 crc kubenswrapper[4948]: E0312 00:06:42.110867 4948 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.180:6443: connect: connection refused" node="crc" Mar 12 00:06:42 crc kubenswrapper[4948]: I0312 00:06:42.242184 4948 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.180:6443: connect: connection refused Mar 12 00:06:42 crc kubenswrapper[4948]: I0312 00:06:42.315563 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"91bc95ec8f59e0db090f0580cb985cb6fb9f01fde5d6de934aa0c096b3c6f280"} Mar 12 00:06:42 crc kubenswrapper[4948]: I0312 00:06:42.316743 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"642dbb27b3c8456c103813e3ea308da2d30af88a3f6ef9a953399ae1a585c7cc"} Mar 12 00:06:42 crc kubenswrapper[4948]: I0312 00:06:42.318122 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"b6d0e1098581c5d959aceffd999a879bdad99e876db181a5c2c00bb1b5dfa153"} Mar 12 00:06:42 crc kubenswrapper[4948]: I0312 00:06:42.319410 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"c7852645f8648eaca077097c34cb29cc20a72599c7c0b4728c01211b97bb57c0"} Mar 12 00:06:42 crc kubenswrapper[4948]: I0312 00:06:42.321896 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"47f8ff230f33182efc550c78b233e370f2a82585c2c8d4d806fbc3d4082cffdc"} Mar 12 00:06:42 crc kubenswrapper[4948]: W0312 00:06:42.469576 4948 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.180:6443: connect: connection refused Mar 12 00:06:42 crc kubenswrapper[4948]: E0312 00:06:42.469873 4948 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.180:6443: connect: connection refused" logger="UnhandledError" Mar 12 00:06:42 crc kubenswrapper[4948]: W0312 00:06:42.603006 4948 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.180:6443: connect: connection refused Mar 12 00:06:42 crc kubenswrapper[4948]: E0312 00:06:42.603082 4948 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.180:6443: connect: connection refused" logger="UnhandledError" Mar 12 00:06:42 crc kubenswrapper[4948]: E0312 00:06:42.649337 4948 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.180:6443: connect: connection refused" interval="1.6s" Mar 12 00:06:42 crc kubenswrapper[4948]: W0312 00:06:42.721654 4948 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": dial tcp 38.102.83.180:6443: connect: connection refused Mar 12 00:06:42 crc kubenswrapper[4948]: E0312 00:06:42.721709 4948 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": dial tcp 38.102.83.180:6443: connect: connection refused" logger="UnhandledError" Mar 12 00:06:42 crc kubenswrapper[4948]: W0312 00:06:42.756085 4948 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": dial tcp 38.102.83.180:6443: connect: connection refused Mar 12 00:06:42 crc kubenswrapper[4948]: E0312 00:06:42.756156 4948 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": dial tcp 38.102.83.180:6443: connect: connection refused" logger="UnhandledError" Mar 12 00:06:42 crc kubenswrapper[4948]: I0312 00:06:42.911682 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:06:42 crc kubenswrapper[4948]: I0312 00:06:42.913585 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:06:42 crc kubenswrapper[4948]: I0312 00:06:42.913634 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:06:42 crc kubenswrapper[4948]: I0312 00:06:42.913651 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:06:42 crc kubenswrapper[4948]: I0312 00:06:42.913683 4948 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 12 00:06:42 crc kubenswrapper[4948]: E0312 00:06:42.914260 4948 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.180:6443: connect: connection refused" node="crc" Mar 12 00:06:43 crc kubenswrapper[4948]: I0312 00:06:43.198147 4948 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 12 00:06:43 crc kubenswrapper[4948]: E0312 00:06:43.200387 4948 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": dial tcp 38.102.83.180:6443: connect: connection refused" logger="UnhandledError" Mar 12 00:06:43 crc kubenswrapper[4948]: I0312 00:06:43.241843 4948 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.180:6443: connect: connection refused Mar 12 00:06:43 crc kubenswrapper[4948]: I0312 00:06:43.326123 4948 generic.go:334] "Generic (PLEG): container finished" podID="d1b160f5dda77d281dd8e69ec8d817f9" containerID="78171d768ed45a0338b479044b9820500c88e17d0d2d9dcbf4b724de06ae3012" exitCode=0 Mar 12 00:06:43 crc kubenswrapper[4948]: I0312 00:06:43.326242 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerDied","Data":"78171d768ed45a0338b479044b9820500c88e17d0d2d9dcbf4b724de06ae3012"} Mar 12 00:06:43 crc kubenswrapper[4948]: I0312 00:06:43.326254 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:06:43 crc kubenswrapper[4948]: I0312 00:06:43.328159 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:06:43 crc kubenswrapper[4948]: I0312 00:06:43.328262 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:06:43 crc kubenswrapper[4948]: I0312 00:06:43.328287 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:06:43 crc kubenswrapper[4948]: I0312 00:06:43.329816 4948 generic.go:334] "Generic (PLEG): container finished" podID="3dcd261975c3d6b9a6ad6367fd4facd3" containerID="c756f1a5b429d17d95df4797725081682b66e9ef3e53a9deb9518e326e40325b" exitCode=0 Mar 12 00:06:43 crc kubenswrapper[4948]: I0312 00:06:43.329924 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:06:43 crc kubenswrapper[4948]: I0312 00:06:43.329920 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerDied","Data":"c756f1a5b429d17d95df4797725081682b66e9ef3e53a9deb9518e326e40325b"} Mar 12 00:06:43 crc kubenswrapper[4948]: I0312 00:06:43.331475 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:06:43 crc kubenswrapper[4948]: I0312 00:06:43.331550 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:06:43 crc kubenswrapper[4948]: I0312 00:06:43.331583 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:06:43 crc kubenswrapper[4948]: I0312 00:06:43.334793 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"663bca87e44109c1d2bb59964e567d8b78fe2e9de39bcec0557cb43e39143add"} Mar 12 00:06:43 crc kubenswrapper[4948]: I0312 00:06:43.334846 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"f580535b16c31fb0ebadbdd9b7851ba2dce2f7d52bcad8c9cacc197f5e7dfad4"} Mar 12 00:06:43 crc kubenswrapper[4948]: I0312 00:06:43.334870 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"fb45482a404d6696cfbc3919997e53146b7b9dd1536056a55374574ae01ff193"} Mar 12 00:06:43 crc kubenswrapper[4948]: I0312 00:06:43.337219 4948 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="7e42f2e94463ac8a3d7980a25122062da9ee87fcdc1f10e260ff9974b2e42359" exitCode=0 Mar 12 00:06:43 crc kubenswrapper[4948]: I0312 00:06:43.337430 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:06:43 crc kubenswrapper[4948]: I0312 00:06:43.337519 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"7e42f2e94463ac8a3d7980a25122062da9ee87fcdc1f10e260ff9974b2e42359"} Mar 12 00:06:43 crc kubenswrapper[4948]: I0312 00:06:43.338732 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:06:43 crc kubenswrapper[4948]: I0312 00:06:43.338780 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:06:43 crc kubenswrapper[4948]: I0312 00:06:43.338797 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:06:43 crc kubenswrapper[4948]: I0312 00:06:43.339553 4948 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="1faa5dc7fd3913910baaefc02adec0696bf70787fdc019624e5f26a7ec343879" exitCode=0 Mar 12 00:06:43 crc kubenswrapper[4948]: I0312 00:06:43.339591 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"1faa5dc7fd3913910baaefc02adec0696bf70787fdc019624e5f26a7ec343879"} Mar 12 00:06:43 crc kubenswrapper[4948]: I0312 00:06:43.339639 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:06:43 crc kubenswrapper[4948]: I0312 00:06:43.340561 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:06:43 crc kubenswrapper[4948]: I0312 00:06:43.340592 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:06:43 crc kubenswrapper[4948]: I0312 00:06:43.341229 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:06:43 crc kubenswrapper[4948]: I0312 00:06:43.341343 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:06:43 crc kubenswrapper[4948]: I0312 00:06:43.346231 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:06:43 crc kubenswrapper[4948]: I0312 00:06:43.346289 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:06:43 crc kubenswrapper[4948]: I0312 00:06:43.346354 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:06:44 crc kubenswrapper[4948]: I0312 00:06:44.242866 4948 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": dial tcp 38.102.83.180:6443: connect: connection refused Mar 12 00:06:44 crc kubenswrapper[4948]: E0312 00:06:44.250571 4948 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.180:6443: connect: connection refused" interval="3.2s" Mar 12 00:06:44 crc kubenswrapper[4948]: I0312 00:06:44.345270 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" event={"ID":"d1b160f5dda77d281dd8e69ec8d817f9","Type":"ContainerStarted","Data":"f58f45d5eb63c29e243da4ed197ef81a00adb088ef9c8b65323dd3ff49f3c5a0"} Mar 12 00:06:44 crc kubenswrapper[4948]: I0312 00:06:44.345417 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:06:44 crc kubenswrapper[4948]: I0312 00:06:44.347174 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:06:44 crc kubenswrapper[4948]: I0312 00:06:44.347204 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:06:44 crc kubenswrapper[4948]: I0312 00:06:44.347217 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:06:44 crc kubenswrapper[4948]: I0312 00:06:44.348801 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"ccb319a44802931d007a0d9265f98bac6f23efee9a7fb82798718a315464e89f"} Mar 12 00:06:44 crc kubenswrapper[4948]: I0312 00:06:44.348847 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"a57137a3fb48ab82c4324c6121b7dbcde8dd33e02cd7012ff42248cc5f164bec"} Mar 12 00:06:44 crc kubenswrapper[4948]: I0312 00:06:44.348868 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" event={"ID":"3dcd261975c3d6b9a6ad6367fd4facd3","Type":"ContainerStarted","Data":"49c25293fe6ddb25b23fc250e3a5d94ce3f2c464f450c0da24dcff2febf15124"} Mar 12 00:06:44 crc kubenswrapper[4948]: I0312 00:06:44.349002 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:06:44 crc kubenswrapper[4948]: I0312 00:06:44.349878 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:06:44 crc kubenswrapper[4948]: I0312 00:06:44.349903 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:06:44 crc kubenswrapper[4948]: I0312 00:06:44.349954 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:06:44 crc kubenswrapper[4948]: I0312 00:06:44.352178 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"0bf5b5d40a0cc76725ae2cf453089b31e75c304cf6037ef06b8ca8f7cff35a5a"} Mar 12 00:06:44 crc kubenswrapper[4948]: I0312 00:06:44.352260 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:06:44 crc kubenswrapper[4948]: I0312 00:06:44.353142 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:06:44 crc kubenswrapper[4948]: I0312 00:06:44.353173 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:06:44 crc kubenswrapper[4948]: I0312 00:06:44.353181 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:06:44 crc kubenswrapper[4948]: I0312 00:06:44.355940 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"d56e32f9cb96ad103bab80d6a730b70d3387264a25a7a61ef1f0726e7290d6c1"} Mar 12 00:06:44 crc kubenswrapper[4948]: I0312 00:06:44.356148 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"1ad13662849131222d5c8ba1041b81fa75bcb1fcbfa2e71c0c751f04b532a5b3"} Mar 12 00:06:44 crc kubenswrapper[4948]: I0312 00:06:44.356172 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"c8cd4faf47b25395ffd42d4bf5416ea29dcaa3ec380d6d6d6a9f13321d35195a"} Mar 12 00:06:44 crc kubenswrapper[4948]: I0312 00:06:44.356192 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"1cb328c8431b39dac735d49726a2eae40a367849b2c146bb2e49d02958f0e5a2"} Mar 12 00:06:44 crc kubenswrapper[4948]: I0312 00:06:44.357656 4948 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="09b8f21b0a3a025179285c54d5afb44216880401be4ff2f891b2687c937df4c5" exitCode=0 Mar 12 00:06:44 crc kubenswrapper[4948]: I0312 00:06:44.357712 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"09b8f21b0a3a025179285c54d5afb44216880401be4ff2f891b2687c937df4c5"} Mar 12 00:06:44 crc kubenswrapper[4948]: I0312 00:06:44.357777 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:06:44 crc kubenswrapper[4948]: I0312 00:06:44.360671 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:06:44 crc kubenswrapper[4948]: I0312 00:06:44.360727 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:06:44 crc kubenswrapper[4948]: I0312 00:06:44.360742 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:06:44 crc kubenswrapper[4948]: W0312 00:06:44.449428 4948 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": dial tcp 38.102.83.180:6443: connect: connection refused Mar 12 00:06:44 crc kubenswrapper[4948]: E0312 00:06:44.449498 4948 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": dial tcp 38.102.83.180:6443: connect: connection refused" logger="UnhandledError" Mar 12 00:06:44 crc kubenswrapper[4948]: I0312 00:06:44.514941 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:06:44 crc kubenswrapper[4948]: I0312 00:06:44.516400 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:06:44 crc kubenswrapper[4948]: I0312 00:06:44.516433 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:06:44 crc kubenswrapper[4948]: I0312 00:06:44.516445 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:06:44 crc kubenswrapper[4948]: I0312 00:06:44.516470 4948 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 12 00:06:44 crc kubenswrapper[4948]: E0312 00:06:44.516949 4948 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": dial tcp 38.102.83.180:6443: connect: connection refused" node="crc" Mar 12 00:06:44 crc kubenswrapper[4948]: W0312 00:06:44.572675 4948 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": dial tcp 38.102.83.180:6443: connect: connection refused Mar 12 00:06:44 crc kubenswrapper[4948]: E0312 00:06:44.572777 4948 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": dial tcp 38.102.83.180:6443: connect: connection refused" logger="UnhandledError" Mar 12 00:06:45 crc kubenswrapper[4948]: I0312 00:06:45.229791 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 12 00:06:45 crc kubenswrapper[4948]: I0312 00:06:45.365376 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"f7757b841decaa29f211c1bc9b62df13b6407d5379bc4f7321f2b506c1de7217"} Mar 12 00:06:45 crc kubenswrapper[4948]: I0312 00:06:45.365503 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:06:45 crc kubenswrapper[4948]: I0312 00:06:45.366795 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:06:45 crc kubenswrapper[4948]: I0312 00:06:45.366849 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:06:45 crc kubenswrapper[4948]: I0312 00:06:45.366867 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:06:45 crc kubenswrapper[4948]: I0312 00:06:45.368215 4948 generic.go:334] "Generic (PLEG): container finished" podID="2139d3e2895fc6797b9c76a1b4c9886d" containerID="521a47b80f4266f2906a0ec020719f0de344f821d79fec1121db683fd7423bca" exitCode=0 Mar 12 00:06:45 crc kubenswrapper[4948]: I0312 00:06:45.368256 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerDied","Data":"521a47b80f4266f2906a0ec020719f0de344f821d79fec1121db683fd7423bca"} Mar 12 00:06:45 crc kubenswrapper[4948]: I0312 00:06:45.368360 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:06:45 crc kubenswrapper[4948]: I0312 00:06:45.368426 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:06:45 crc kubenswrapper[4948]: I0312 00:06:45.368524 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 12 00:06:45 crc kubenswrapper[4948]: I0312 00:06:45.368432 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:06:45 crc kubenswrapper[4948]: I0312 00:06:45.369397 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:06:45 crc kubenswrapper[4948]: I0312 00:06:45.369608 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:06:45 crc kubenswrapper[4948]: I0312 00:06:45.369650 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:06:45 crc kubenswrapper[4948]: I0312 00:06:45.369668 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:06:45 crc kubenswrapper[4948]: I0312 00:06:45.369703 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:06:45 crc kubenswrapper[4948]: I0312 00:06:45.369735 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:06:45 crc kubenswrapper[4948]: I0312 00:06:45.369751 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:06:45 crc kubenswrapper[4948]: I0312 00:06:45.370206 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:06:45 crc kubenswrapper[4948]: I0312 00:06:45.370255 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:06:45 crc kubenswrapper[4948]: I0312 00:06:45.370273 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:06:45 crc kubenswrapper[4948]: I0312 00:06:45.370807 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:06:45 crc kubenswrapper[4948]: I0312 00:06:45.370848 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:06:45 crc kubenswrapper[4948]: I0312 00:06:45.370866 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:06:45 crc kubenswrapper[4948]: I0312 00:06:45.599588 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 00:06:46 crc kubenswrapper[4948]: I0312 00:06:46.377993 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"b67015b6fc7fd6c325413c690fea22cb50e306ea073c56dbc125d42bd188ddc1"} Mar 12 00:06:46 crc kubenswrapper[4948]: I0312 00:06:46.378090 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"44c61a7725acb6f6e42f445dc00889061bbb39e9769205b56e97586f87c5dcec"} Mar 12 00:06:46 crc kubenswrapper[4948]: I0312 00:06:46.378115 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"14ee53252e073893911a486ddb3f0fad62fa4312cf69004a16aa4489e9847261"} Mar 12 00:06:46 crc kubenswrapper[4948]: I0312 00:06:46.378185 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:06:46 crc kubenswrapper[4948]: I0312 00:06:46.378217 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:06:46 crc kubenswrapper[4948]: I0312 00:06:46.378334 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:06:46 crc kubenswrapper[4948]: I0312 00:06:46.378358 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 00:06:46 crc kubenswrapper[4948]: I0312 00:06:46.381602 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:06:46 crc kubenswrapper[4948]: I0312 00:06:46.381618 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:06:46 crc kubenswrapper[4948]: I0312 00:06:46.381642 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:06:46 crc kubenswrapper[4948]: I0312 00:06:46.381662 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:06:46 crc kubenswrapper[4948]: I0312 00:06:46.381672 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:06:46 crc kubenswrapper[4948]: I0312 00:06:46.381698 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:06:46 crc kubenswrapper[4948]: I0312 00:06:46.381618 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:06:46 crc kubenswrapper[4948]: I0312 00:06:46.381767 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:06:46 crc kubenswrapper[4948]: I0312 00:06:46.381794 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:06:47 crc kubenswrapper[4948]: I0312 00:06:47.257931 4948 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 12 00:06:47 crc kubenswrapper[4948]: I0312 00:06:47.387580 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"32a7e838e79a3814f01067dad935dc0ceadb2e1243535aca2e716c08ccf76e5c"} Mar 12 00:06:47 crc kubenswrapper[4948]: I0312 00:06:47.388293 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd/etcd-crc" event={"ID":"2139d3e2895fc6797b9c76a1b4c9886d","Type":"ContainerStarted","Data":"f499a58bf5d7dfac37625579269383f728921324b0be860626eab788cdbadb6b"} Mar 12 00:06:47 crc kubenswrapper[4948]: I0312 00:06:47.387693 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:06:47 crc kubenswrapper[4948]: I0312 00:06:47.387670 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:06:47 crc kubenswrapper[4948]: I0312 00:06:47.390219 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:06:47 crc kubenswrapper[4948]: I0312 00:06:47.390257 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:06:47 crc kubenswrapper[4948]: I0312 00:06:47.390274 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:06:47 crc kubenswrapper[4948]: I0312 00:06:47.390630 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:06:47 crc kubenswrapper[4948]: I0312 00:06:47.390788 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:06:47 crc kubenswrapper[4948]: I0312 00:06:47.390925 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:06:47 crc kubenswrapper[4948]: I0312 00:06:47.718058 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:06:47 crc kubenswrapper[4948]: I0312 00:06:47.719923 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:06:47 crc kubenswrapper[4948]: I0312 00:06:47.719992 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:06:47 crc kubenswrapper[4948]: I0312 00:06:47.720013 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:06:47 crc kubenswrapper[4948]: I0312 00:06:47.720052 4948 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 12 00:06:48 crc kubenswrapper[4948]: I0312 00:06:48.230260 4948 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 12 00:06:48 crc kubenswrapper[4948]: I0312 00:06:48.230375 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers)" Mar 12 00:06:48 crc kubenswrapper[4948]: I0312 00:06:48.389890 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:06:48 crc kubenswrapper[4948]: I0312 00:06:48.390950 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:06:48 crc kubenswrapper[4948]: I0312 00:06:48.391005 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:06:48 crc kubenswrapper[4948]: I0312 00:06:48.391023 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:06:49 crc kubenswrapper[4948]: I0312 00:06:49.402355 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 00:06:49 crc kubenswrapper[4948]: I0312 00:06:49.402612 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:06:49 crc kubenswrapper[4948]: I0312 00:06:49.404022 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:06:49 crc kubenswrapper[4948]: I0312 00:06:49.404076 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:06:49 crc kubenswrapper[4948]: I0312 00:06:49.404101 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:06:51 crc kubenswrapper[4948]: I0312 00:06:51.362228 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-etcd/etcd-crc" Mar 12 00:06:51 crc kubenswrapper[4948]: I0312 00:06:51.362565 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:06:51 crc kubenswrapper[4948]: I0312 00:06:51.364049 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:06:51 crc kubenswrapper[4948]: I0312 00:06:51.364112 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:06:51 crc kubenswrapper[4948]: I0312 00:06:51.364147 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:06:51 crc kubenswrapper[4948]: E0312 00:06:51.409752 4948 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 12 00:06:51 crc kubenswrapper[4948]: I0312 00:06:51.615090 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 12 00:06:51 crc kubenswrapper[4948]: I0312 00:06:51.615286 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:06:51 crc kubenswrapper[4948]: I0312 00:06:51.616765 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:06:51 crc kubenswrapper[4948]: I0312 00:06:51.616813 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:06:51 crc kubenswrapper[4948]: I0312 00:06:51.616834 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:06:51 crc kubenswrapper[4948]: I0312 00:06:51.775591 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-etcd/etcd-crc" Mar 12 00:06:51 crc kubenswrapper[4948]: I0312 00:06:51.775765 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:06:51 crc kubenswrapper[4948]: I0312 00:06:51.777164 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:06:51 crc kubenswrapper[4948]: I0312 00:06:51.777225 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:06:51 crc kubenswrapper[4948]: I0312 00:06:51.777242 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:06:51 crc kubenswrapper[4948]: I0312 00:06:51.878524 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 12 00:06:52 crc kubenswrapper[4948]: I0312 00:06:52.402687 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:06:52 crc kubenswrapper[4948]: I0312 00:06:52.404477 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:06:52 crc kubenswrapper[4948]: I0312 00:06:52.404539 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:06:52 crc kubenswrapper[4948]: I0312 00:06:52.404566 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:06:53 crc kubenswrapper[4948]: I0312 00:06:53.221168 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 12 00:06:53 crc kubenswrapper[4948]: I0312 00:06:53.228190 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 12 00:06:53 crc kubenswrapper[4948]: I0312 00:06:53.405416 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:06:53 crc kubenswrapper[4948]: I0312 00:06:53.406710 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:06:53 crc kubenswrapper[4948]: I0312 00:06:53.406762 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:06:53 crc kubenswrapper[4948]: I0312 00:06:53.406780 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:06:53 crc kubenswrapper[4948]: I0312 00:06:53.413663 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 12 00:06:54 crc kubenswrapper[4948]: I0312 00:06:54.407855 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:06:54 crc kubenswrapper[4948]: I0312 00:06:54.409082 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:06:54 crc kubenswrapper[4948]: I0312 00:06:54.409135 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:06:54 crc kubenswrapper[4948]: I0312 00:06:54.409154 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:06:54 crc kubenswrapper[4948]: W0312 00:06:54.842599 4948 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout Mar 12 00:06:54 crc kubenswrapper[4948]: I0312 00:06:54.842723 4948 trace.go:236] Trace[62789529]: "Reflector ListAndWatch" name:k8s.io/client-go/informers/factory.go:160 (12-Mar-2026 00:06:44.841) (total time: 10001ms): Mar 12 00:06:54 crc kubenswrapper[4948]: Trace[62789529]: ---"Objects listed" error:Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": net/http: TLS handshake timeout 10001ms (00:06:54.842) Mar 12 00:06:54 crc kubenswrapper[4948]: Trace[62789529]: [10.001414406s] [10.001414406s] END Mar 12 00:06:54 crc kubenswrapper[4948]: E0312 00:06:54.842755 4948 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": net/http: TLS handshake timeout" logger="UnhandledError" Mar 12 00:06:55 crc kubenswrapper[4948]: I0312 00:06:55.243115 4948 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": net/http: TLS handshake timeout Mar 12 00:06:55 crc kubenswrapper[4948]: I0312 00:06:55.410278 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:06:55 crc kubenswrapper[4948]: I0312 00:06:55.411526 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:06:55 crc kubenswrapper[4948]: I0312 00:06:55.411735 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:06:55 crc kubenswrapper[4948]: I0312 00:06:55.411922 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:06:55 crc kubenswrapper[4948]: E0312 00:06:55.756724 4948 certificate_manager.go:562] "Unhandled Error" err="kubernetes.io/kube-apiserver-client-kubelet: Failed while requesting a signed certificate from the control plane: cannot create certificate signing request: Post \"https://api-int.crc.testing:6443/apis/certificates.k8s.io/v1/certificatesigningrequests\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:06:55Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 12 00:06:55 crc kubenswrapper[4948]: E0312 00:06:55.760876 4948 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:06:55Z is after 2026-02-23T05:33:13Z" interval="6.4s" Mar 12 00:06:55 crc kubenswrapper[4948]: E0312 00:06:55.765901 4948 kubelet_node_status.go:99] "Unable to register node with API server" err="Post \"https://api-int.crc.testing:6443/api/v1/nodes\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:06:55Z is after 2026-02-23T05:33:13Z" node="crc" Mar 12 00:06:55 crc kubenswrapper[4948]: W0312 00:06:55.768411 4948 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:06:55Z is after 2026-02-23T05:33:13Z Mar 12 00:06:55 crc kubenswrapper[4948]: E0312 00:06:55.768519 4948 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:06:55Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 12 00:06:55 crc kubenswrapper[4948]: W0312 00:06:55.771903 4948 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: Get "https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:06:55Z is after 2026-02-23T05:33:13Z Mar 12 00:06:55 crc kubenswrapper[4948]: E0312 00:06:55.772153 4948 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: Get \"https://api-int.crc.testing:6443/api/v1/services?fieldSelector=spec.clusterIP%21%3DNone&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:06:55Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 12 00:06:55 crc kubenswrapper[4948]: E0312 00:06:55.777030 4948 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/default/events\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:06:55Z is after 2026-02-23T05:33:13Z" event="&Event{ObjectMeta:{crc.189bef4f15559a92 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:41.239808658 +0000 UTC m=+0.695412426,LastTimestamp:2026-03-12 00:06:41.239808658 +0000 UTC m=+0.695412426,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:06:55 crc kubenswrapper[4948]: W0312 00:06:55.784173 4948 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: Get "https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:06:55Z is after 2026-02-23T05:33:13Z Mar 12 00:06:55 crc kubenswrapper[4948]: E0312 00:06:55.784279 4948 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: Get \"https://api-int.crc.testing:6443/apis/node.k8s.io/v1/runtimeclasses?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:06:55Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 12 00:06:55 crc kubenswrapper[4948]: I0312 00:06:55.786873 4948 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 12 00:06:55 crc kubenswrapper[4948]: I0312 00:06:55.786945 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Mar 12 00:06:55 crc kubenswrapper[4948]: I0312 00:06:55.792690 4948 patch_prober.go:28] interesting pod/kube-apiserver-crc container/kube-apiserver namespace/openshift-kube-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 403" start-of-body={"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 12 00:06:55 crc kubenswrapper[4948]: I0312 00:06:55.792747 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 403" Mar 12 00:06:56 crc kubenswrapper[4948]: I0312 00:06:56.247659 4948 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:06:56Z is after 2026-02-23T05:33:13Z Mar 12 00:06:56 crc kubenswrapper[4948]: I0312 00:06:56.414756 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Mar 12 00:06:56 crc kubenswrapper[4948]: I0312 00:06:56.418734 4948 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="f7757b841decaa29f211c1bc9b62df13b6407d5379bc4f7321f2b506c1de7217" exitCode=255 Mar 12 00:06:56 crc kubenswrapper[4948]: I0312 00:06:56.418813 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"f7757b841decaa29f211c1bc9b62df13b6407d5379bc4f7321f2b506c1de7217"} Mar 12 00:06:56 crc kubenswrapper[4948]: I0312 00:06:56.419253 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:06:56 crc kubenswrapper[4948]: I0312 00:06:56.420451 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:06:56 crc kubenswrapper[4948]: I0312 00:06:56.420484 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:06:56 crc kubenswrapper[4948]: I0312 00:06:56.420503 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:06:56 crc kubenswrapper[4948]: I0312 00:06:56.425198 4948 scope.go:117] "RemoveContainer" containerID="f7757b841decaa29f211c1bc9b62df13b6407d5379bc4f7321f2b506c1de7217" Mar 12 00:06:57 crc kubenswrapper[4948]: I0312 00:06:57.245621 4948 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:06:57Z is after 2026-02-23T05:33:13Z Mar 12 00:06:57 crc kubenswrapper[4948]: I0312 00:06:57.423516 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Mar 12 00:06:57 crc kubenswrapper[4948]: I0312 00:06:57.425722 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"040118c06a9c12cb485fc6f741cddb66d181d83621e241212afd2a1d46bfc910"} Mar 12 00:06:57 crc kubenswrapper[4948]: I0312 00:06:57.425910 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:06:57 crc kubenswrapper[4948]: I0312 00:06:57.427059 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:06:57 crc kubenswrapper[4948]: I0312 00:06:57.427100 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:06:57 crc kubenswrapper[4948]: I0312 00:06:57.427118 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:06:58 crc kubenswrapper[4948]: I0312 00:06:58.231131 4948 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 12 00:06:58 crc kubenswrapper[4948]: I0312 00:06:58.231191 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 12 00:06:58 crc kubenswrapper[4948]: I0312 00:06:58.246561 4948 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:06:58Z is after 2026-02-23T05:33:13Z Mar 12 00:06:58 crc kubenswrapper[4948]: W0312 00:06:58.381998 4948 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: Get "https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:06:58Z is after 2026-02-23T05:33:13Z Mar 12 00:06:58 crc kubenswrapper[4948]: E0312 00:06:58.382182 4948 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: Get \"https://api-int.crc.testing:6443/api/v1/nodes?fieldSelector=metadata.name%3Dcrc&limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:06:58Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 12 00:06:58 crc kubenswrapper[4948]: I0312 00:06:58.430061 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 12 00:06:58 crc kubenswrapper[4948]: I0312 00:06:58.430795 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/0.log" Mar 12 00:06:58 crc kubenswrapper[4948]: I0312 00:06:58.432918 4948 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="040118c06a9c12cb485fc6f741cddb66d181d83621e241212afd2a1d46bfc910" exitCode=255 Mar 12 00:06:58 crc kubenswrapper[4948]: I0312 00:06:58.432965 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"040118c06a9c12cb485fc6f741cddb66d181d83621e241212afd2a1d46bfc910"} Mar 12 00:06:58 crc kubenswrapper[4948]: I0312 00:06:58.433031 4948 scope.go:117] "RemoveContainer" containerID="f7757b841decaa29f211c1bc9b62df13b6407d5379bc4f7321f2b506c1de7217" Mar 12 00:06:58 crc kubenswrapper[4948]: I0312 00:06:58.433347 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:06:58 crc kubenswrapper[4948]: I0312 00:06:58.434871 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:06:58 crc kubenswrapper[4948]: I0312 00:06:58.434901 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:06:58 crc kubenswrapper[4948]: I0312 00:06:58.434911 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:06:58 crc kubenswrapper[4948]: I0312 00:06:58.435504 4948 scope.go:117] "RemoveContainer" containerID="040118c06a9c12cb485fc6f741cddb66d181d83621e241212afd2a1d46bfc910" Mar 12 00:06:58 crc kubenswrapper[4948]: E0312 00:06:58.435676 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 12 00:06:59 crc kubenswrapper[4948]: I0312 00:06:59.245072 4948 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:06:59Z is after 2026-02-23T05:33:13Z Mar 12 00:06:59 crc kubenswrapper[4948]: W0312 00:06:59.246517 4948 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:06:59Z is after 2026-02-23T05:33:13Z Mar 12 00:06:59 crc kubenswrapper[4948]: E0312 00:06:59.246611 4948 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: Get \"https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csidrivers?limit=500&resourceVersion=0\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:06:59Z is after 2026-02-23T05:33:13Z" logger="UnhandledError" Mar 12 00:06:59 crc kubenswrapper[4948]: I0312 00:06:59.410668 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 00:06:59 crc kubenswrapper[4948]: I0312 00:06:59.439035 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 12 00:06:59 crc kubenswrapper[4948]: I0312 00:06:59.441834 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:06:59 crc kubenswrapper[4948]: I0312 00:06:59.443378 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:06:59 crc kubenswrapper[4948]: I0312 00:06:59.443635 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:06:59 crc kubenswrapper[4948]: I0312 00:06:59.443821 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:06:59 crc kubenswrapper[4948]: I0312 00:06:59.444925 4948 scope.go:117] "RemoveContainer" containerID="040118c06a9c12cb485fc6f741cddb66d181d83621e241212afd2a1d46bfc910" Mar 12 00:06:59 crc kubenswrapper[4948]: E0312 00:06:59.445389 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 12 00:06:59 crc kubenswrapper[4948]: I0312 00:06:59.450247 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 00:07:00 crc kubenswrapper[4948]: I0312 00:07:00.247450 4948 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:07:00Z is after 2026-02-23T05:33:13Z Mar 12 00:07:00 crc kubenswrapper[4948]: I0312 00:07:00.444867 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:07:00 crc kubenswrapper[4948]: I0312 00:07:00.446654 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:07:00 crc kubenswrapper[4948]: I0312 00:07:00.446708 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:07:00 crc kubenswrapper[4948]: I0312 00:07:00.446728 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:07:00 crc kubenswrapper[4948]: I0312 00:07:00.447800 4948 scope.go:117] "RemoveContainer" containerID="040118c06a9c12cb485fc6f741cddb66d181d83621e241212afd2a1d46bfc910" Mar 12 00:07:00 crc kubenswrapper[4948]: E0312 00:07:00.448133 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 12 00:07:00 crc kubenswrapper[4948]: I0312 00:07:00.897806 4948 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 00:07:01 crc kubenswrapper[4948]: I0312 00:07:01.250803 4948 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: Get "https://api-int.crc.testing:6443/apis/storage.k8s.io/v1/csinodes/crc?resourceVersion=0": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:07:01Z is after 2026-02-23T05:33:13Z Mar 12 00:07:01 crc kubenswrapper[4948]: E0312 00:07:01.409932 4948 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 12 00:07:01 crc kubenswrapper[4948]: I0312 00:07:01.452201 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:07:01 crc kubenswrapper[4948]: I0312 00:07:01.453472 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:07:01 crc kubenswrapper[4948]: I0312 00:07:01.453526 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:07:01 crc kubenswrapper[4948]: I0312 00:07:01.453552 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:07:01 crc kubenswrapper[4948]: I0312 00:07:01.454466 4948 scope.go:117] "RemoveContainer" containerID="040118c06a9c12cb485fc6f741cddb66d181d83621e241212afd2a1d46bfc910" Mar 12 00:07:01 crc kubenswrapper[4948]: E0312 00:07:01.454741 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 12 00:07:01 crc kubenswrapper[4948]: I0312 00:07:01.817793 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-etcd/etcd-crc" Mar 12 00:07:01 crc kubenswrapper[4948]: I0312 00:07:01.818016 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:07:01 crc kubenswrapper[4948]: I0312 00:07:01.819376 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:07:01 crc kubenswrapper[4948]: I0312 00:07:01.819482 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:07:01 crc kubenswrapper[4948]: I0312 00:07:01.819512 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:07:01 crc kubenswrapper[4948]: I0312 00:07:01.840924 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-etcd/etcd-crc" Mar 12 00:07:02 crc kubenswrapper[4948]: I0312 00:07:02.167140 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:07:02 crc kubenswrapper[4948]: I0312 00:07:02.168924 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:07:02 crc kubenswrapper[4948]: I0312 00:07:02.168967 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:07:02 crc kubenswrapper[4948]: I0312 00:07:02.168985 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:07:02 crc kubenswrapper[4948]: I0312 00:07:02.169018 4948 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 12 00:07:02 crc kubenswrapper[4948]: E0312 00:07:02.169148 4948 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 12 00:07:02 crc kubenswrapper[4948]: E0312 00:07:02.173082 4948 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 12 00:07:02 crc kubenswrapper[4948]: I0312 00:07:02.248784 4948 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 00:07:02 crc kubenswrapper[4948]: I0312 00:07:02.455472 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:07:02 crc kubenswrapper[4948]: I0312 00:07:02.456941 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:07:02 crc kubenswrapper[4948]: I0312 00:07:02.457031 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:07:02 crc kubenswrapper[4948]: I0312 00:07:02.457050 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:07:02 crc kubenswrapper[4948]: W0312 00:07:02.484355 4948 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User "system:anonymous" cannot list resource "runtimeclasses" in API group "node.k8s.io" at the cluster scope Mar 12 00:07:02 crc kubenswrapper[4948]: E0312 00:07:02.484474 4948 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"runtimeclasses\" in API group \"node.k8s.io\" at the cluster scope" logger="UnhandledError" Mar 12 00:07:03 crc kubenswrapper[4948]: I0312 00:07:03.248428 4948 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 00:07:03 crc kubenswrapper[4948]: I0312 00:07:03.975329 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 00:07:03 crc kubenswrapper[4948]: I0312 00:07:03.975559 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:07:03 crc kubenswrapper[4948]: I0312 00:07:03.976949 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:07:03 crc kubenswrapper[4948]: I0312 00:07:03.977016 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:07:03 crc kubenswrapper[4948]: I0312 00:07:03.977035 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:07:03 crc kubenswrapper[4948]: I0312 00:07:03.977841 4948 scope.go:117] "RemoveContainer" containerID="040118c06a9c12cb485fc6f741cddb66d181d83621e241212afd2a1d46bfc910" Mar 12 00:07:03 crc kubenswrapper[4948]: E0312 00:07:03.978145 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 12 00:07:04 crc kubenswrapper[4948]: W0312 00:07:04.092843 4948 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: services is forbidden: User "system:anonymous" cannot list resource "services" in API group "" at the cluster scope Mar 12 00:07:04 crc kubenswrapper[4948]: E0312 00:07:04.092919 4948 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: services is forbidden: User \"system:anonymous\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError" Mar 12 00:07:04 crc kubenswrapper[4948]: I0312 00:07:04.248358 4948 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 00:07:04 crc kubenswrapper[4948]: I0312 00:07:04.352404 4948 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Rotating certificates Mar 12 00:07:04 crc kubenswrapper[4948]: I0312 00:07:04.373446 4948 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Mar 12 00:07:05 crc kubenswrapper[4948]: I0312 00:07:05.247391 4948 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 00:07:05 crc kubenswrapper[4948]: E0312 00:07:05.786253 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189bef4f15559a92 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:Starting,Message:Starting kubelet.,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:41.239808658 +0000 UTC m=+0.695412426,LastTimestamp:2026-03-12 00:06:41.239808658 +0000 UTC m=+0.695412426,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:05 crc kubenswrapper[4948]: E0312 00:07:05.790762 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189bef4f19783a8d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:41.309186701 +0000 UTC m=+0.764790479,LastTimestamp:2026-03-12 00:06:41.309186701 +0000 UTC m=+0.764790479,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:05 crc kubenswrapper[4948]: E0312 00:07:05.796981 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189bef4f197b73eb default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:41.309397995 +0000 UTC m=+0.765001763,LastTimestamp:2026-03-12 00:06:41.309397995 +0000 UTC m=+0.765001763,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:05 crc kubenswrapper[4948]: E0312 00:07:05.806237 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189bef4f197c6aee default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:41.30946123 +0000 UTC m=+0.765064998,LastTimestamp:2026-03-12 00:06:41.30946123 +0000 UTC m=+0.765064998,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:05 crc kubenswrapper[4948]: E0312 00:07:05.813409 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189bef4f1f6f6385 default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeAllocatableEnforced,Message:Updated Node Allocatable limit across pods,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:41.409270661 +0000 UTC m=+0.864874439,LastTimestamp:2026-03-12 00:06:41.409270661 +0000 UTC m=+0.864874439,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:05 crc kubenswrapper[4948]: E0312 00:07:05.820591 4948 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189bef4f19783a8d\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189bef4f19783a8d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:41.309186701 +0000 UTC m=+0.764790479,LastTimestamp:2026-03-12 00:06:41.41307204 +0000 UTC m=+0.868675818,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:05 crc kubenswrapper[4948]: E0312 00:07:05.827352 4948 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189bef4f197b73eb\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189bef4f197b73eb default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:41.309397995 +0000 UTC m=+0.765001763,LastTimestamp:2026-03-12 00:06:41.413099358 +0000 UTC m=+0.868703136,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:05 crc kubenswrapper[4948]: E0312 00:07:05.834970 4948 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189bef4f197c6aee\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189bef4f197c6aee default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:41.30946123 +0000 UTC m=+0.765064998,LastTimestamp:2026-03-12 00:06:41.413119606 +0000 UTC m=+0.868723394,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:05 crc kubenswrapper[4948]: E0312 00:07:05.842593 4948 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189bef4f19783a8d\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189bef4f19783a8d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:41.309186701 +0000 UTC m=+0.764790479,LastTimestamp:2026-03-12 00:06:41.414395165 +0000 UTC m=+0.869998943,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:05 crc kubenswrapper[4948]: E0312 00:07:05.847583 4948 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189bef4f197b73eb\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189bef4f197b73eb default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:41.309397995 +0000 UTC m=+0.765001763,LastTimestamp:2026-03-12 00:06:41.414428382 +0000 UTC m=+0.870032160,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:05 crc kubenswrapper[4948]: E0312 00:07:05.853767 4948 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189bef4f197c6aee\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189bef4f197c6aee default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:41.30946123 +0000 UTC m=+0.765064998,LastTimestamp:2026-03-12 00:06:41.414444781 +0000 UTC m=+0.870048549,Count:3,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:05 crc kubenswrapper[4948]: E0312 00:07:05.861518 4948 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189bef4f19783a8d\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189bef4f19783a8d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:41.309186701 +0000 UTC m=+0.764790479,LastTimestamp:2026-03-12 00:06:41.415132187 +0000 UTC m=+0.870735955,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:05 crc kubenswrapper[4948]: E0312 00:07:05.868617 4948 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189bef4f197b73eb\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189bef4f197b73eb default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:41.309397995 +0000 UTC m=+0.765001763,LastTimestamp:2026-03-12 00:06:41.415297374 +0000 UTC m=+0.870901152,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:05 crc kubenswrapper[4948]: E0312 00:07:05.873817 4948 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189bef4f197c6aee\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189bef4f197c6aee default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:41.30946123 +0000 UTC m=+0.765064998,LastTimestamp:2026-03-12 00:06:41.415398676 +0000 UTC m=+0.871002454,Count:4,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:05 crc kubenswrapper[4948]: E0312 00:07:05.879255 4948 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189bef4f19783a8d\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189bef4f19783a8d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:41.309186701 +0000 UTC m=+0.764790479,LastTimestamp:2026-03-12 00:06:41.417634658 +0000 UTC m=+0.873238426,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:05 crc kubenswrapper[4948]: E0312 00:07:05.883890 4948 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189bef4f197b73eb\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189bef4f197b73eb default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:41.309397995 +0000 UTC m=+0.765001763,LastTimestamp:2026-03-12 00:06:41.417656476 +0000 UTC m=+0.873260244,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:05 crc kubenswrapper[4948]: E0312 00:07:05.888977 4948 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189bef4f19783a8d\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189bef4f19783a8d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:41.309186701 +0000 UTC m=+0.764790479,LastTimestamp:2026-03-12 00:06:41.417667835 +0000 UTC m=+0.873271573,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:05 crc kubenswrapper[4948]: E0312 00:07:05.895176 4948 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189bef4f197c6aee\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189bef4f197c6aee default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:41.30946123 +0000 UTC m=+0.765064998,LastTimestamp:2026-03-12 00:06:41.417673365 +0000 UTC m=+0.873277143,Count:5,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:05 crc kubenswrapper[4948]: E0312 00:07:05.899421 4948 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189bef4f197b73eb\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189bef4f197b73eb default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:41.309397995 +0000 UTC m=+0.765001763,LastTimestamp:2026-03-12 00:06:41.417683354 +0000 UTC m=+0.873287092,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:05 crc kubenswrapper[4948]: E0312 00:07:05.903405 4948 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189bef4f197c6aee\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189bef4f197c6aee default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:41.30946123 +0000 UTC m=+0.765064998,LastTimestamp:2026-03-12 00:06:41.417694963 +0000 UTC m=+0.873298701,Count:6,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:05 crc kubenswrapper[4948]: E0312 00:07:05.907849 4948 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189bef4f19783a8d\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189bef4f19783a8d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:41.309186701 +0000 UTC m=+0.764790479,LastTimestamp:2026-03-12 00:06:41.419182315 +0000 UTC m=+0.874786053,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:05 crc kubenswrapper[4948]: E0312 00:07:05.913827 4948 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189bef4f197b73eb\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189bef4f197b73eb default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:41.309397995 +0000 UTC m=+0.765001763,LastTimestamp:2026-03-12 00:06:41.419200294 +0000 UTC m=+0.874804032,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:05 crc kubenswrapper[4948]: E0312 00:07:05.918389 4948 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189bef4f197c6aee\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189bef4f197c6aee default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientPID,Message:Node crc status is now: NodeHasSufficientPID,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:41.30946123 +0000 UTC m=+0.765064998,LastTimestamp:2026-03-12 00:06:41.419211713 +0000 UTC m=+0.874815451,Count:7,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:05 crc kubenswrapper[4948]: E0312 00:07:05.922895 4948 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189bef4f19783a8d\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189bef4f19783a8d default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasSufficientMemory,Message:Node crc status is now: NodeHasSufficientMemory,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:41.309186701 +0000 UTC m=+0.764790479,LastTimestamp:2026-03-12 00:06:41.422497943 +0000 UTC m=+0.878101681,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:05 crc kubenswrapper[4948]: E0312 00:07:05.929215 4948 event.go:359] "Server rejected event (will not retry!)" err="events \"crc.189bef4f197b73eb\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"default\"" event="&Event{ObjectMeta:{crc.189bef4f197b73eb default 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Node,Namespace:,Name:crc,UID:crc,APIVersion:,ResourceVersion:,FieldPath:,},Reason:NodeHasNoDiskPressure,Message:Node crc status is now: NodeHasNoDiskPressure,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:41.309397995 +0000 UTC m=+0.765001763,LastTimestamp:2026-03-12 00:06:41.422521261 +0000 UTC m=+0.878124999,Count:8,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:05 crc kubenswrapper[4948]: E0312 00:07:05.937374 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189bef4f3958c53d openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:41.843995965 +0000 UTC m=+1.299599723,LastTimestamp:2026-03-12 00:06:41.843995965 +0000 UTC m=+1.299599723,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:05 crc kubenswrapper[4948]: E0312 00:07:05.943735 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189bef4f39599ff9 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:41.844051961 +0000 UTC m=+1.299655699,LastTimestamp:2026-03-12 00:06:41.844051961 +0000 UTC m=+1.299655699,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:05 crc kubenswrapper[4948]: E0312 00:07:05.950193 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189bef4f39dcdb0b openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:41.852652299 +0000 UTC m=+1.308256047,LastTimestamp:2026-03-12 00:06:41.852652299 +0000 UTC m=+1.308256047,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:05 crc kubenswrapper[4948]: E0312 00:07:05.955938 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189bef4f3ab2ea88 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:41.866680968 +0000 UTC m=+1.322284716,LastTimestamp:2026-03-12 00:06:41.866680968 +0000 UTC m=+1.322284716,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:05 crc kubenswrapper[4948]: E0312 00:07:05.961810 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189bef4f3ab9f795 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:41.867143061 +0000 UTC m=+1.322746839,LastTimestamp:2026-03-12 00:06:41.867143061 +0000 UTC m=+1.322746839,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:05 crc kubenswrapper[4948]: E0312 00:07:05.967732 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189bef4f5f7d6ddf openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Created,Message:Created container wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:42.483932639 +0000 UTC m=+1.939536387,LastTimestamp:2026-03-12 00:06:42.483932639 +0000 UTC m=+1.939536387,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:05 crc kubenswrapper[4948]: E0312 00:07:05.972824 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189bef4f5f7ed2ea openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Created,Message:Created container kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:42.484024042 +0000 UTC m=+1.939627780,LastTimestamp:2026-03-12 00:06:42.484024042 +0000 UTC m=+1.939627780,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:05 crc kubenswrapper[4948]: E0312 00:07:05.981933 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189bef4f5f8616a6 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:42.484500134 +0000 UTC m=+1.940103872,LastTimestamp:2026-03-12 00:06:42.484500134 +0000 UTC m=+1.940103872,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:05 crc kubenswrapper[4948]: E0312 00:07:05.989060 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189bef4f5f9d7484 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:42.486031492 +0000 UTC m=+1.941635230,LastTimestamp:2026-03-12 00:06:42.486031492 +0000 UTC m=+1.941635230,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:05 crc kubenswrapper[4948]: E0312 00:07:05.995721 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189bef4f5fb8e1f5 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Created,Message:Created container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:42.487828981 +0000 UTC m=+1.943432719,LastTimestamp:2026-03-12 00:06:42.487828981 +0000 UTC m=+1.943432719,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:05 crc kubenswrapper[4948]: E0312 00:07:05.999069 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189bef4f60550228 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{wait-for-host-port},},Reason:Started,Message:Started container wait-for-host-port,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:42.49806084 +0000 UTC m=+1.953664578,LastTimestamp:2026-03-12 00:06:42.49806084 +0000 UTC m=+1.953664578,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.001465 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189bef4f60613077 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager},},Reason:Started,Message:Started container kube-controller-manager,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:42.498859127 +0000 UTC m=+1.954462875,LastTimestamp:2026-03-12 00:06:42.498859127 +0000 UTC m=+1.954462875,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.004617 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189bef4f6075faa9 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:42.500221609 +0000 UTC m=+1.955825347,LastTimestamp:2026-03-12 00:06:42.500221609 +0000 UTC m=+1.955825347,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.006415 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189bef4f609d723b openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:42.502808123 +0000 UTC m=+1.958411871,LastTimestamp:2026-03-12 00:06:42.502808123 +0000 UTC m=+1.958411871,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.010764 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189bef4f60ba1c99 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:42.504686745 +0000 UTC m=+1.960290493,LastTimestamp:2026-03-12 00:06:42.504686745 +0000 UTC m=+1.960290493,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.016741 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189bef4f60d57afe openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{setup},},Reason:Started,Message:Started container setup,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:42.506480382 +0000 UTC m=+1.962084120,LastTimestamp:2026-03-12 00:06:42.506480382 +0000 UTC m=+1.962084120,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.023123 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189bef4f73473e49 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:42.815925833 +0000 UTC m=+2.271529571,LastTimestamp:2026-03-12 00:06:42.815925833 +0000 UTC m=+2.271529571,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.027295 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189bef4f73e952dc openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:42.826547932 +0000 UTC m=+2.282151670,LastTimestamp:2026-03-12 00:06:42.826547932 +0000 UTC m=+2.282151670,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.033943 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189bef4f7401884c openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:42.828134476 +0000 UTC m=+2.283738254,LastTimestamp:2026-03-12 00:06:42.828134476 +0000 UTC m=+2.283738254,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.038969 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189bef4f81b440eb openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Created,Message:Created container kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:43.057950955 +0000 UTC m=+2.513554733,LastTimestamp:2026-03-12 00:06:43.057950955 +0000 UTC m=+2.513554733,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.045498 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189bef4f824c07d2 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-cert-syncer},},Reason:Started,Message:Started container kube-controller-manager-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:43.06789781 +0000 UTC m=+2.523501548,LastTimestamp:2026-03-12 00:06:43.06789781 +0000 UTC m=+2.523501548,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.052538 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189bef4f82664e2b openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:43.069619755 +0000 UTC m=+2.525223533,LastTimestamp:2026-03-12 00:06:43.069619755 +0000 UTC m=+2.525223533,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.060179 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189bef4f91eb1a90 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:43.329981072 +0000 UTC m=+2.785584840,LastTimestamp:2026-03-12 00:06:43.329981072 +0000 UTC m=+2.785584840,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.067992 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189bef4f921331c2 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Created,Message:Created container kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:43.33260845 +0000 UTC m=+2.788212218,LastTimestamp:2026-03-12 00:06:43.33260845 +0000 UTC m=+2.788212218,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.074504 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189bef4f921f7935 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:43.333413173 +0000 UTC m=+2.789016921,LastTimestamp:2026-03-12 00:06:43.333413173 +0000 UTC m=+2.789016921,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.081490 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189bef4f9287e6af openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:43.340256943 +0000 UTC m=+2.795860711,LastTimestamp:2026-03-12 00:06:43.340256943 +0000 UTC m=+2.795860711,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.089174 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189bef4f92f3da70 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-controller-manager-recovery-controller},},Reason:Started,Message:Started container kube-controller-manager-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:43.347331696 +0000 UTC m=+2.802935444,LastTimestamp:2026-03-12 00:06:43.347331696 +0000 UTC m=+2.802935444,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.097703 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189bef4f934ba6be openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:43.35308563 +0000 UTC m=+2.808689368,LastTimestamp:2026-03-12 00:06:43.35308563 +0000 UTC m=+2.808689368,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.103937 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189bef4fa095568a openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Created,Message:Created container kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:43.57601857 +0000 UTC m=+3.031622308,LastTimestamp:2026-03-12 00:06:43.57601857 +0000 UTC m=+3.031622308,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.110862 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189bef4fa0b075a5 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Created,Message:Created container kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:43.577796005 +0000 UTC m=+3.033399743,LastTimestamp:2026-03-12 00:06:43.577796005 +0000 UTC m=+3.033399743,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.125478 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189bef4fa0f360c1 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Created,Message:Created container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:43.582181569 +0000 UTC m=+3.037785307,LastTimestamp:2026-03-12 00:06:43.582181569 +0000 UTC m=+3.037785307,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.134770 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189bef4fa160466e openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Created,Message:Created container etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:43.589318254 +0000 UTC m=+3.044921992,LastTimestamp:2026-03-12 00:06:43.589318254 +0000 UTC m=+3.044921992,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.142045 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189bef4fa18c4e65 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler},},Reason:Started,Message:Started container kube-scheduler,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:43.592203877 +0000 UTC m=+3.047807635,LastTimestamp:2026-03-12 00:06:43.592203877 +0000 UTC m=+3.047807635,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.151005 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189bef4fa1986a2a openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Started,Message:Started container kube-apiserver,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:43.592997418 +0000 UTC m=+3.048601156,LastTimestamp:2026-03-12 00:06:43.592997418 +0000 UTC m=+3.048601156,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.158457 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189bef4fa1a4d208 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:43.59381044 +0000 UTC m=+3.049414188,LastTimestamp:2026-03-12 00:06:43.59381044 +0000 UTC m=+3.049414188,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.167445 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189bef4fa1a57f29 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:43.593854761 +0000 UTC m=+3.049458499,LastTimestamp:2026-03-12 00:06:43.593854761 +0000 UTC m=+3.049458499,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.176231 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-machine-config-operator\"" event="&Event{ObjectMeta:{kube-rbac-proxy-crio-crc.189bef4fa1dca393 openshift-machine-config-operator 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-machine-config-operator,Name:kube-rbac-proxy-crio-crc,UID:d1b160f5dda77d281dd8e69ec8d817f9,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-rbac-proxy-crio},},Reason:Started,Message:Started container kube-rbac-proxy-crio,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:43.597468563 +0000 UTC m=+3.053072321,LastTimestamp:2026-03-12 00:06:43.597468563 +0000 UTC m=+3.053072321,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.183815 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189bef4fa3023644 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-ensure-env-vars},},Reason:Started,Message:Started container etcd-ensure-env-vars,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:43.616708164 +0000 UTC m=+3.072311902,LastTimestamp:2026-03-12 00:06:43.616708164 +0000 UTC m=+3.072311902,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.190070 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189bef4fad10801c openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Created,Message:Created container kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:43.785416732 +0000 UTC m=+3.241020480,LastTimestamp:2026-03-12 00:06:43.785416732 +0000 UTC m=+3.241020480,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.199132 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189bef4fad109f3e openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Created,Message:Created container kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:43.785424702 +0000 UTC m=+3.241028440,LastTimestamp:2026-03-12 00:06:43.785424702 +0000 UTC m=+3.241028440,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.207517 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189bef4fadbef9b9 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-cert-syncer},},Reason:Started,Message:Started container kube-scheduler-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:43.796851129 +0000 UTC m=+3.252454867,LastTimestamp:2026-03-12 00:06:43.796851129 +0000 UTC m=+3.252454867,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.213127 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189bef4fadcc5ecb openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:43.797728971 +0000 UTC m=+3.253332709,LastTimestamp:2026-03-12 00:06:43.797728971 +0000 UTC m=+3.253332709,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.220887 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189bef4fadde826b openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-syncer},},Reason:Started,Message:Started container kube-apiserver-cert-syncer,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:43.798917739 +0000 UTC m=+3.254521467,LastTimestamp:2026-03-12 00:06:43.798917739 +0000 UTC m=+3.254521467,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.229073 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189bef4fade95655 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:43.799627349 +0000 UTC m=+3.255231087,LastTimestamp:2026-03-12 00:06:43.799627349 +0000 UTC m=+3.255231087,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.236253 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189bef4fb9ed7b2d openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Created,Message:Created container kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:44.001225517 +0000 UTC m=+3.456829265,LastTimestamp:2026-03-12 00:06:44.001225517 +0000 UTC m=+3.456829265,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: I0312 00:07:06.243391 4948 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.244606 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189bef4fb9f656fb openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Created,Message:Created container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:44.001806075 +0000 UTC m=+3.457409813,LastTimestamp:2026-03-12 00:06:44.001806075 +0000 UTC m=+3.457409813,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.250450 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-scheduler\"" event="&Event{ObjectMeta:{openshift-kube-scheduler-crc.189bef4fbaf8b907 openshift-kube-scheduler 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-scheduler,Name:openshift-kube-scheduler-crc,UID:3dcd261975c3d6b9a6ad6367fd4facd3,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-scheduler-recovery-controller},},Reason:Started,Message:Started container kube-scheduler-recovery-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:44.018739463 +0000 UTC m=+3.474343201,LastTimestamp:2026-03-12 00:06:44.018739463 +0000 UTC m=+3.474343201,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.256525 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189bef4fbb0b1a86 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-cert-regeneration-controller},},Reason:Started,Message:Started container kube-apiserver-cert-regeneration-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:44.01994407 +0000 UTC m=+3.475547808,LastTimestamp:2026-03-12 00:06:44.01994407 +0000 UTC m=+3.475547808,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.261520 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189bef4fbb19b181 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:44.020900225 +0000 UTC m=+3.476503963,LastTimestamp:2026-03-12 00:06:44.020900225 +0000 UTC m=+3.476503963,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.268093 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189bef4fc64fad8e openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Created,Message:Created container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:44.208987534 +0000 UTC m=+3.664591272,LastTimestamp:2026-03-12 00:06:44.208987534 +0000 UTC m=+3.664591272,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.274851 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189bef4fc719ce07 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-insecure-readyz},},Reason:Started,Message:Started container kube-apiserver-insecure-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:44.222234119 +0000 UTC m=+3.677837897,LastTimestamp:2026-03-12 00:06:44.222234119 +0000 UTC m=+3.677837897,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.279442 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189bef4fc72c17c5 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:44.223432645 +0000 UTC m=+3.679036383,LastTimestamp:2026-03-12 00:06:44.223432645 +0000 UTC m=+3.679036383,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.287774 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189bef4fcf7a888d openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:44.362791053 +0000 UTC m=+3.818394801,LastTimestamp:2026-03-12 00:06:44.362791053 +0000 UTC m=+3.818394801,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.294734 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189bef4fd3a7800a openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Created,Message:Created container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:44.432846858 +0000 UTC m=+3.888450596,LastTimestamp:2026-03-12 00:06:44.432846858 +0000 UTC m=+3.888450596,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.301435 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189bef4fd40d91e0 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Started,Message:Started container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:44.439536096 +0000 UTC m=+3.895139834,LastTimestamp:2026-03-12 00:06:44.439536096 +0000 UTC m=+3.895139834,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.309198 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189bef4fdbb5c469 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Created,Message:Created container etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:44.567999593 +0000 UTC m=+4.023603341,LastTimestamp:2026-03-12 00:06:44.567999593 +0000 UTC m=+4.023603341,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.316594 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189bef4fdc954b66 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.initContainers{etcd-resources-copy},},Reason:Started,Message:Started container etcd-resources-copy,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:44.582648678 +0000 UTC m=+4.038252436,LastTimestamp:2026-03-12 00:06:44.582648678 +0000 UTC m=+4.038252436,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.326552 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189bef500bae502b openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:45.372817451 +0000 UTC m=+4.828421219,LastTimestamp:2026-03-12 00:06:45.372817451 +0000 UTC m=+4.828421219,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.333670 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189bef501b28d87b openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Created,Message:Created container etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:45.632505979 +0000 UTC m=+5.088109757,LastTimestamp:2026-03-12 00:06:45.632505979 +0000 UTC m=+5.088109757,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.341644 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189bef501bfcf60e openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcdctl},},Reason:Started,Message:Started container etcdctl,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:45.646407182 +0000 UTC m=+5.102010960,LastTimestamp:2026-03-12 00:06:45.646407182 +0000 UTC m=+5.102010960,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.349177 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189bef501c11fa0e openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:45.647784462 +0000 UTC m=+5.103388240,LastTimestamp:2026-03-12 00:06:45.647784462 +0000 UTC m=+5.103388240,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.356513 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189bef502b80d04c openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Created,Message:Created container etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:45.906706508 +0000 UTC m=+5.362310276,LastTimestamp:2026-03-12 00:06:45.906706508 +0000 UTC m=+5.362310276,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.364089 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189bef502c4b20f7 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd},},Reason:Started,Message:Started container etcd,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:45.919965431 +0000 UTC m=+5.375569199,LastTimestamp:2026-03-12 00:06:45.919965431 +0000 UTC m=+5.375569199,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.371494 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189bef502c597971 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:45.920905585 +0000 UTC m=+5.376509353,LastTimestamp:2026-03-12 00:06:45.920905585 +0000 UTC m=+5.376509353,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.378566 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189bef503b750030 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Created,Message:Created container etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:46.174367792 +0000 UTC m=+5.629971560,LastTimestamp:2026-03-12 00:06:46.174367792 +0000 UTC m=+5.629971560,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.385993 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189bef503c4ff01d openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-metrics},},Reason:Started,Message:Started container etcd-metrics,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:46.188716061 +0000 UTC m=+5.644319839,LastTimestamp:2026-03-12 00:06:46.188716061 +0000 UTC m=+5.644319839,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.392652 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189bef503c6972ad openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:46.190387885 +0000 UTC m=+5.645991663,LastTimestamp:2026-03-12 00:06:46.190387885 +0000 UTC m=+5.645991663,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.400882 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189bef504bd03d0c openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Created,Message:Created container etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:46.448782604 +0000 UTC m=+5.904386372,LastTimestamp:2026-03-12 00:06:46.448782604 +0000 UTC m=+5.904386372,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.408129 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189bef504cb83636 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-readyz},},Reason:Started,Message:Started container etcd-readyz,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:46.463985206 +0000 UTC m=+5.919588974,LastTimestamp:2026-03-12 00:06:46.463985206 +0000 UTC m=+5.919588974,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.415378 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189bef504cca37b1 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:46.465165233 +0000 UTC m=+5.920769011,LastTimestamp:2026-03-12 00:06:46.465165233 +0000 UTC m=+5.920769011,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.419273 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189bef505976079f openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Created,Message:Created container etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:46.677751711 +0000 UTC m=+6.133355449,LastTimestamp:2026-03-12 00:06:46.677751711 +0000 UTC m=+6.133355449,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.424794 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-etcd\"" event="&Event{ObjectMeta:{etcd-crc.189bef505a6f7771 openshift-etcd 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-etcd,Name:etcd-crc,UID:2139d3e2895fc6797b9c76a1b4c9886d,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{etcd-rev},},Reason:Started,Message:Started container etcd-rev,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:46.694098801 +0000 UTC m=+6.149702569,LastTimestamp:2026-03-12 00:06:46.694098801 +0000 UTC m=+6.149702569,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.433063 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 12 00:07:06 crc kubenswrapper[4948]: &Event{ObjectMeta:{kube-controller-manager-crc.189bef50b600a6a7 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": context deadline exceeded (Client.Timeout exceeded while awaiting headers) Mar 12 00:07:06 crc kubenswrapper[4948]: body: Mar 12 00:07:06 crc kubenswrapper[4948]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:48.230340263 +0000 UTC m=+7.685944031,LastTimestamp:2026-03-12 00:06:48.230340263 +0000 UTC m=+7.685944031,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 12 00:07:06 crc kubenswrapper[4948]: > Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.439797 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189bef50b601af3a openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": context deadline exceeded (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:48.230407994 +0000 UTC m=+7.686011732,LastTimestamp:2026-03-12 00:06:48.230407994 +0000 UTC m=+7.686011732,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.447210 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 12 00:07:06 crc kubenswrapper[4948]: &Event{ObjectMeta:{kube-apiserver-crc.189bef527868f72e openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 403 Mar 12 00:07:06 crc kubenswrapper[4948]: body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 12 00:07:06 crc kubenswrapper[4948]: Mar 12 00:07:06 crc kubenswrapper[4948]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:55.786923822 +0000 UTC m=+15.242527600,LastTimestamp:2026-03-12 00:06:55.786923822 +0000 UTC m=+15.242527600,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 12 00:07:06 crc kubenswrapper[4948]: > Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.453652 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189bef527869be70 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 403,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:55.786974832 +0000 UTC m=+15.242578610,LastTimestamp:2026-03-12 00:06:55.786974832 +0000 UTC m=+15.242578610,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.459782 4948 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189bef527868f72e\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event=< Mar 12 00:07:06 crc kubenswrapper[4948]: &Event{ObjectMeta:{kube-apiserver-crc.189bef527868f72e openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:ProbeError,Message:Startup probe error: HTTP probe failed with statuscode: 403 Mar 12 00:07:06 crc kubenswrapper[4948]: body: {"kind":"Status","apiVersion":"v1","metadata":{},"status":"Failure","message":"forbidden: User \"system:anonymous\" cannot get path \"/livez\"","reason":"Forbidden","details":{},"code":403} Mar 12 00:07:06 crc kubenswrapper[4948]: Mar 12 00:07:06 crc kubenswrapper[4948]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:55.786923822 +0000 UTC m=+15.242527600,LastTimestamp:2026-03-12 00:06:55.792730766 +0000 UTC m=+15.248334514,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 12 00:07:06 crc kubenswrapper[4948]: > Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.463977 4948 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189bef527869be70\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189bef527869be70 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver},},Reason:Unhealthy,Message:Startup probe failed: HTTP probe failed with statuscode: 403,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:55.786974832 +0000 UTC m=+15.242578610,LastTimestamp:2026-03-12 00:06:55.792766607 +0000 UTC m=+15.248370355,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.469525 4948 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189bef4fc72c17c5\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189bef4fc72c17c5 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:44.223432645 +0000 UTC m=+3.679036383,LastTimestamp:2026-03-12 00:06:56.438418208 +0000 UTC m=+15.894021956,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.476949 4948 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189bef4fd3a7800a\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189bef4fd3a7800a openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Created,Message:Created container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:44.432846858 +0000 UTC m=+3.888450596,LastTimestamp:2026-03-12 00:06:56.674996237 +0000 UTC m=+16.130600005,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.483524 4948 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-apiserver-crc.189bef4fd40d91e0\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-apiserver\"" event="&Event{ObjectMeta:{kube-apiserver-crc.189bef4fd40d91e0 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-crc,UID:f4b27818a5e8e43d0dc095d08835c792,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{kube-apiserver-check-endpoints},},Reason:Started,Message:Started container kube-apiserver-check-endpoints,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:44.439536096 +0000 UTC m=+3.895139834,LastTimestamp:2026-03-12 00:06:56.686663718 +0000 UTC m=+16.142267486,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.489204 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 12 00:07:06 crc kubenswrapper[4948]: &Event{ObjectMeta:{kube-controller-manager-crc.189bef530a19446b openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 12 00:07:06 crc kubenswrapper[4948]: body: Mar 12 00:07:06 crc kubenswrapper[4948]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:58.231174251 +0000 UTC m=+17.686777989,LastTimestamp:2026-03-12 00:06:58.231174251 +0000 UTC m=+17.686777989,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 12 00:07:06 crc kubenswrapper[4948]: > Mar 12 00:07:06 crc kubenswrapper[4948]: E0312 00:07:06.496081 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189bef530a19e03d openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:58.231214141 +0000 UTC m=+17.686817879,LastTimestamp:2026-03-12 00:06:58.231214141 +0000 UTC m=+17.686817879,Count:1,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:07 crc kubenswrapper[4948]: I0312 00:07:07.246469 4948 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 00:07:08 crc kubenswrapper[4948]: I0312 00:07:08.230773 4948 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/cluster-policy-controller namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 12 00:07:08 crc kubenswrapper[4948]: I0312 00:07:08.231012 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" probeResult="failure" output="Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 12 00:07:08 crc kubenswrapper[4948]: I0312 00:07:08.231090 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 12 00:07:08 crc kubenswrapper[4948]: I0312 00:07:08.231278 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:07:08 crc kubenswrapper[4948]: I0312 00:07:08.232932 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:07:08 crc kubenswrapper[4948]: I0312 00:07:08.232985 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:07:08 crc kubenswrapper[4948]: I0312 00:07:08.233002 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:07:08 crc kubenswrapper[4948]: I0312 00:07:08.233873 4948 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="cluster-policy-controller" containerStatusID={"Type":"cri-o","ID":"f580535b16c31fb0ebadbdd9b7851ba2dce2f7d52bcad8c9cacc197f5e7dfad4"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container cluster-policy-controller failed startup probe, will be restarted" Mar 12 00:07:08 crc kubenswrapper[4948]: I0312 00:07:08.234139 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="cluster-policy-controller" containerID="cri-o://f580535b16c31fb0ebadbdd9b7851ba2dce2f7d52bcad8c9cacc197f5e7dfad4" gracePeriod=30 Mar 12 00:07:08 crc kubenswrapper[4948]: E0312 00:07:08.236671 4948 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189bef530a19446b\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event=< Mar 12 00:07:08 crc kubenswrapper[4948]: &Event{ObjectMeta:{kube-controller-manager-crc.189bef530a19446b openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:ProbeError,Message:Startup probe error: Get "https://192.168.126.11:10357/healthz": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers) Mar 12 00:07:08 crc kubenswrapper[4948]: body: Mar 12 00:07:08 crc kubenswrapper[4948]: ,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:58.231174251 +0000 UTC m=+17.686777989,LastTimestamp:2026-03-12 00:07:08.230936575 +0000 UTC m=+27.686540343,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,} Mar 12 00:07:08 crc kubenswrapper[4948]: > Mar 12 00:07:08 crc kubenswrapper[4948]: E0312 00:07:08.244442 4948 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189bef530a19e03d\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189bef530a19e03d openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Unhealthy,Message:Startup probe failed: Get \"https://192.168.126.11:10357/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers),Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:58.231214141 +0000 UTC m=+17.686817879,LastTimestamp:2026-03-12 00:07:08.231051888 +0000 UTC m=+27.686655666,Count:2,Type:Warning,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:08 crc kubenswrapper[4948]: I0312 00:07:08.244663 4948 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 00:07:08 crc kubenswrapper[4948]: E0312 00:07:08.246810 4948 event.go:359] "Server rejected event (will not retry!)" err="events is forbidden: User \"system:anonymous\" cannot create resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189bef555e520441 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Killing,Message:Container cluster-policy-controller failed startup probe, will be restarted,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:07:08.234114113 +0000 UTC m=+27.689717891,LastTimestamp:2026-03-12 00:07:08.234114113 +0000 UTC m=+27.689717891,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:08 crc kubenswrapper[4948]: E0312 00:07:08.376563 4948 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189bef4f6075faa9\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189bef4f6075faa9 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Pulled,Message:Container image \"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:42.500221609 +0000 UTC m=+1.955825347,LastTimestamp:2026-03-12 00:07:08.365291094 +0000 UTC m=+27.820894862,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:08 crc kubenswrapper[4948]: I0312 00:07:08.479116 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 12 00:07:08 crc kubenswrapper[4948]: I0312 00:07:08.479832 4948 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="f580535b16c31fb0ebadbdd9b7851ba2dce2f7d52bcad8c9cacc197f5e7dfad4" exitCode=255 Mar 12 00:07:08 crc kubenswrapper[4948]: I0312 00:07:08.479922 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"f580535b16c31fb0ebadbdd9b7851ba2dce2f7d52bcad8c9cacc197f5e7dfad4"} Mar 12 00:07:09 crc kubenswrapper[4948]: I0312 00:07:09.173261 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:07:09 crc kubenswrapper[4948]: I0312 00:07:09.175231 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:07:09 crc kubenswrapper[4948]: I0312 00:07:09.175557 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:07:09 crc kubenswrapper[4948]: I0312 00:07:09.176137 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:07:09 crc kubenswrapper[4948]: I0312 00:07:09.176232 4948 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 12 00:07:09 crc kubenswrapper[4948]: E0312 00:07:09.179602 4948 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 12 00:07:09 crc kubenswrapper[4948]: E0312 00:07:09.180059 4948 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 12 00:07:10 crc kubenswrapper[4948]: I0312 00:07:10.263648 4948 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 00:07:10 crc kubenswrapper[4948]: E0312 00:07:10.314553 4948 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189bef4f73473e49\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189bef4f73473e49 openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Created,Message:Created container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:42.815925833 +0000 UTC m=+2.271529571,LastTimestamp:2026-03-12 00:07:10.306044668 +0000 UTC m=+29.761648446,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:10 crc kubenswrapper[4948]: E0312 00:07:10.325621 4948 event.go:359] "Server rejected event (will not retry!)" err="events \"kube-controller-manager-crc.189bef4f73e952dc\" is forbidden: User \"system:anonymous\" cannot patch resource \"events\" in API group \"\" in the namespace \"openshift-kube-controller-manager\"" event="&Event{ObjectMeta:{kube-controller-manager-crc.189bef4f73e952dc openshift-kube-controller-manager 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-controller-manager,Name:kube-controller-manager-crc,UID:f614b9022728cf315e60c057852e563e,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{cluster-policy-controller},},Reason:Started,Message:Started container cluster-policy-controller,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:06:42.826547932 +0000 UTC m=+2.282151670,LastTimestamp:2026-03-12 00:07:10.319781799 +0000 UTC m=+29.775385567,Count:2,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:07:11 crc kubenswrapper[4948]: W0312 00:07:11.126644 4948 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: nodes "crc" is forbidden: User "system:anonymous" cannot list resource "nodes" in API group "" at the cluster scope Mar 12 00:07:11 crc kubenswrapper[4948]: E0312 00:07:11.126716 4948 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: nodes \"crc\" is forbidden: User \"system:anonymous\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError" Mar 12 00:07:11 crc kubenswrapper[4948]: W0312 00:07:11.202555 4948 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:anonymous" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope Mar 12 00:07:11 crc kubenswrapper[4948]: E0312 00:07:11.202622 4948 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" Mar 12 00:07:11 crc kubenswrapper[4948]: I0312 00:07:11.249016 4948 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 00:07:11 crc kubenswrapper[4948]: I0312 00:07:11.264403 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 12 00:07:11 crc kubenswrapper[4948]: I0312 00:07:11.265017 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"dc9760bd6ed0d092081fdd2ad33dda2d67f1305531640a35b9dce84b5d86a813"} Mar 12 00:07:11 crc kubenswrapper[4948]: I0312 00:07:11.265124 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:07:11 crc kubenswrapper[4948]: I0312 00:07:11.266383 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:07:11 crc kubenswrapper[4948]: I0312 00:07:11.266436 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:07:11 crc kubenswrapper[4948]: I0312 00:07:11.266455 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:07:11 crc kubenswrapper[4948]: E0312 00:07:11.410148 4948 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 12 00:07:11 crc kubenswrapper[4948]: I0312 00:07:11.616136 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 12 00:07:12 crc kubenswrapper[4948]: I0312 00:07:12.249427 4948 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 00:07:12 crc kubenswrapper[4948]: I0312 00:07:12.267705 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:07:12 crc kubenswrapper[4948]: I0312 00:07:12.268965 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:07:12 crc kubenswrapper[4948]: I0312 00:07:12.269157 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:07:12 crc kubenswrapper[4948]: I0312 00:07:12.269284 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:07:13 crc kubenswrapper[4948]: I0312 00:07:13.249576 4948 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 00:07:13 crc kubenswrapper[4948]: I0312 00:07:13.269961 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:07:13 crc kubenswrapper[4948]: I0312 00:07:13.271280 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:07:13 crc kubenswrapper[4948]: I0312 00:07:13.271436 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:07:13 crc kubenswrapper[4948]: I0312 00:07:13.271453 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:07:14 crc kubenswrapper[4948]: I0312 00:07:14.248963 4948 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 00:07:15 crc kubenswrapper[4948]: I0312 00:07:15.230916 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 12 00:07:15 crc kubenswrapper[4948]: I0312 00:07:15.231999 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:07:15 crc kubenswrapper[4948]: I0312 00:07:15.233659 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:07:15 crc kubenswrapper[4948]: I0312 00:07:15.233719 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:07:15 crc kubenswrapper[4948]: I0312 00:07:15.233733 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:07:15 crc kubenswrapper[4948]: I0312 00:07:15.242864 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 12 00:07:15 crc kubenswrapper[4948]: I0312 00:07:15.248787 4948 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 00:07:15 crc kubenswrapper[4948]: I0312 00:07:15.275514 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:07:15 crc kubenswrapper[4948]: I0312 00:07:15.276673 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:07:15 crc kubenswrapper[4948]: I0312 00:07:15.276726 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:07:15 crc kubenswrapper[4948]: I0312 00:07:15.276745 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:07:16 crc kubenswrapper[4948]: I0312 00:07:16.180560 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:07:16 crc kubenswrapper[4948]: I0312 00:07:16.182552 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:07:16 crc kubenswrapper[4948]: I0312 00:07:16.182611 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:07:16 crc kubenswrapper[4948]: I0312 00:07:16.182632 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:07:16 crc kubenswrapper[4948]: I0312 00:07:16.182672 4948 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 12 00:07:16 crc kubenswrapper[4948]: E0312 00:07:16.187843 4948 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 12 00:07:16 crc kubenswrapper[4948]: E0312 00:07:16.188436 4948 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 12 00:07:16 crc kubenswrapper[4948]: I0312 00:07:16.248473 4948 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 00:07:17 crc kubenswrapper[4948]: I0312 00:07:17.252149 4948 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 00:07:18 crc kubenswrapper[4948]: I0312 00:07:18.248656 4948 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 00:07:18 crc kubenswrapper[4948]: I0312 00:07:18.312603 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:07:18 crc kubenswrapper[4948]: I0312 00:07:18.313958 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:07:18 crc kubenswrapper[4948]: I0312 00:07:18.314180 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:07:18 crc kubenswrapper[4948]: I0312 00:07:18.314379 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:07:18 crc kubenswrapper[4948]: I0312 00:07:18.315359 4948 scope.go:117] "RemoveContainer" containerID="040118c06a9c12cb485fc6f741cddb66d181d83621e241212afd2a1d46bfc910" Mar 12 00:07:19 crc kubenswrapper[4948]: I0312 00:07:19.247257 4948 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 00:07:19 crc kubenswrapper[4948]: I0312 00:07:19.286455 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 12 00:07:19 crc kubenswrapper[4948]: I0312 00:07:19.287791 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"e89f42aa5b3aa0882190e7501b56ba6f208e6c796e0b4041be478564c71b1c87"} Mar 12 00:07:19 crc kubenswrapper[4948]: I0312 00:07:19.287932 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:07:19 crc kubenswrapper[4948]: I0312 00:07:19.288603 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:07:19 crc kubenswrapper[4948]: I0312 00:07:19.288634 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:07:19 crc kubenswrapper[4948]: I0312 00:07:19.288644 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:07:20 crc kubenswrapper[4948]: I0312 00:07:20.249748 4948 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 00:07:20 crc kubenswrapper[4948]: I0312 00:07:20.293559 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 12 00:07:20 crc kubenswrapper[4948]: I0312 00:07:20.294045 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/1.log" Mar 12 00:07:20 crc kubenswrapper[4948]: I0312 00:07:20.296609 4948 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="e89f42aa5b3aa0882190e7501b56ba6f208e6c796e0b4041be478564c71b1c87" exitCode=255 Mar 12 00:07:20 crc kubenswrapper[4948]: I0312 00:07:20.296680 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"e89f42aa5b3aa0882190e7501b56ba6f208e6c796e0b4041be478564c71b1c87"} Mar 12 00:07:20 crc kubenswrapper[4948]: I0312 00:07:20.296744 4948 scope.go:117] "RemoveContainer" containerID="040118c06a9c12cb485fc6f741cddb66d181d83621e241212afd2a1d46bfc910" Mar 12 00:07:20 crc kubenswrapper[4948]: I0312 00:07:20.296925 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:07:20 crc kubenswrapper[4948]: I0312 00:07:20.298039 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:07:20 crc kubenswrapper[4948]: I0312 00:07:20.298078 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:07:20 crc kubenswrapper[4948]: I0312 00:07:20.298095 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:07:20 crc kubenswrapper[4948]: I0312 00:07:20.298844 4948 scope.go:117] "RemoveContainer" containerID="e89f42aa5b3aa0882190e7501b56ba6f208e6c796e0b4041be478564c71b1c87" Mar 12 00:07:20 crc kubenswrapper[4948]: E0312 00:07:20.299044 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 12 00:07:20 crc kubenswrapper[4948]: I0312 00:07:20.896986 4948 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 00:07:21 crc kubenswrapper[4948]: I0312 00:07:21.249116 4948 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 00:07:21 crc kubenswrapper[4948]: I0312 00:07:21.302517 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 12 00:07:21 crc kubenswrapper[4948]: I0312 00:07:21.304934 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:07:21 crc kubenswrapper[4948]: I0312 00:07:21.306104 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:07:21 crc kubenswrapper[4948]: I0312 00:07:21.306144 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:07:21 crc kubenswrapper[4948]: I0312 00:07:21.306160 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:07:21 crc kubenswrapper[4948]: I0312 00:07:21.306923 4948 scope.go:117] "RemoveContainer" containerID="e89f42aa5b3aa0882190e7501b56ba6f208e6c796e0b4041be478564c71b1c87" Mar 12 00:07:21 crc kubenswrapper[4948]: E0312 00:07:21.307190 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 12 00:07:21 crc kubenswrapper[4948]: E0312 00:07:21.411170 4948 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 12 00:07:21 crc kubenswrapper[4948]: I0312 00:07:21.621919 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 12 00:07:21 crc kubenswrapper[4948]: I0312 00:07:21.622118 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:07:21 crc kubenswrapper[4948]: I0312 00:07:21.623602 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:07:21 crc kubenswrapper[4948]: I0312 00:07:21.623644 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:07:21 crc kubenswrapper[4948]: I0312 00:07:21.623660 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:07:22 crc kubenswrapper[4948]: I0312 00:07:22.247485 4948 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 00:07:23 crc kubenswrapper[4948]: I0312 00:07:23.189061 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:07:23 crc kubenswrapper[4948]: I0312 00:07:23.190801 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:07:23 crc kubenswrapper[4948]: I0312 00:07:23.190875 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:07:23 crc kubenswrapper[4948]: I0312 00:07:23.190893 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:07:23 crc kubenswrapper[4948]: I0312 00:07:23.190934 4948 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 12 00:07:23 crc kubenswrapper[4948]: E0312 00:07:23.193984 4948 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 12 00:07:23 crc kubenswrapper[4948]: E0312 00:07:23.194272 4948 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 12 00:07:23 crc kubenswrapper[4948]: I0312 00:07:23.245119 4948 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 00:07:23 crc kubenswrapper[4948]: W0312 00:07:23.521605 4948 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User "system:anonymous" cannot list resource "runtimeclasses" in API group "node.k8s.io" at the cluster scope Mar 12 00:07:23 crc kubenswrapper[4948]: E0312 00:07:23.521672 4948 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.RuntimeClass: failed to list *v1.RuntimeClass: runtimeclasses.node.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"runtimeclasses\" in API group \"node.k8s.io\" at the cluster scope" logger="UnhandledError" Mar 12 00:07:23 crc kubenswrapper[4948]: I0312 00:07:23.974447 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 00:07:23 crc kubenswrapper[4948]: I0312 00:07:23.975381 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:07:23 crc kubenswrapper[4948]: I0312 00:07:23.976834 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:07:23 crc kubenswrapper[4948]: I0312 00:07:23.976897 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:07:23 crc kubenswrapper[4948]: I0312 00:07:23.976922 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:07:23 crc kubenswrapper[4948]: I0312 00:07:23.978061 4948 scope.go:117] "RemoveContainer" containerID="e89f42aa5b3aa0882190e7501b56ba6f208e6c796e0b4041be478564c71b1c87" Mar 12 00:07:23 crc kubenswrapper[4948]: E0312 00:07:23.978465 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 12 00:07:24 crc kubenswrapper[4948]: I0312 00:07:24.250514 4948 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 00:07:24 crc kubenswrapper[4948]: W0312 00:07:24.282538 4948 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Service: services is forbidden: User "system:anonymous" cannot list resource "services" in API group "" at the cluster scope Mar 12 00:07:24 crc kubenswrapper[4948]: E0312 00:07:24.282626 4948 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Service: failed to list *v1.Service: services is forbidden: User \"system:anonymous\" cannot list resource \"services\" in API group \"\" at the cluster scope" logger="UnhandledError" Mar 12 00:07:25 crc kubenswrapper[4948]: I0312 00:07:25.248449 4948 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 00:07:26 crc kubenswrapper[4948]: I0312 00:07:26.249553 4948 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 00:07:26 crc kubenswrapper[4948]: W0312 00:07:26.475369 4948 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.Node: nodes "crc" is forbidden: User "system:anonymous" cannot list resource "nodes" in API group "" at the cluster scope Mar 12 00:07:26 crc kubenswrapper[4948]: E0312 00:07:26.475443 4948 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.Node: failed to list *v1.Node: nodes \"crc\" is forbidden: User \"system:anonymous\" cannot list resource \"nodes\" in API group \"\" at the cluster scope" logger="UnhandledError" Mar 12 00:07:27 crc kubenswrapper[4948]: I0312 00:07:27.250752 4948 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 00:07:28 crc kubenswrapper[4948]: I0312 00:07:28.248497 4948 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 00:07:29 crc kubenswrapper[4948]: I0312 00:07:29.246123 4948 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 00:07:29 crc kubenswrapper[4948]: W0312 00:07:29.807278 4948 reflector.go:561] k8s.io/client-go/informers/factory.go:160: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User "system:anonymous" cannot list resource "csidrivers" in API group "storage.k8s.io" at the cluster scope Mar 12 00:07:29 crc kubenswrapper[4948]: E0312 00:07:29.807404 4948 reflector.go:158] "Unhandled Error" err="k8s.io/client-go/informers/factory.go:160: Failed to watch *v1.CSIDriver: failed to list *v1.CSIDriver: csidrivers.storage.k8s.io is forbidden: User \"system:anonymous\" cannot list resource \"csidrivers\" in API group \"storage.k8s.io\" at the cluster scope" logger="UnhandledError" Mar 12 00:07:30 crc kubenswrapper[4948]: I0312 00:07:30.194383 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:07:30 crc kubenswrapper[4948]: I0312 00:07:30.195803 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:07:30 crc kubenswrapper[4948]: I0312 00:07:30.195859 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:07:30 crc kubenswrapper[4948]: I0312 00:07:30.195877 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:07:30 crc kubenswrapper[4948]: I0312 00:07:30.195915 4948 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 12 00:07:30 crc kubenswrapper[4948]: E0312 00:07:30.201548 4948 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 12 00:07:30 crc kubenswrapper[4948]: E0312 00:07:30.202285 4948 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 12 00:07:30 crc kubenswrapper[4948]: I0312 00:07:30.243116 4948 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 00:07:31 crc kubenswrapper[4948]: I0312 00:07:31.246387 4948 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 00:07:31 crc kubenswrapper[4948]: E0312 00:07:31.412365 4948 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 12 00:07:31 crc kubenswrapper[4948]: I0312 00:07:31.889792 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" Mar 12 00:07:31 crc kubenswrapper[4948]: I0312 00:07:31.890094 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:07:31 crc kubenswrapper[4948]: I0312 00:07:31.891722 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:07:31 crc kubenswrapper[4948]: I0312 00:07:31.891766 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:07:31 crc kubenswrapper[4948]: I0312 00:07:31.891776 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:07:32 crc kubenswrapper[4948]: I0312 00:07:32.247964 4948 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 00:07:33 crc kubenswrapper[4948]: I0312 00:07:33.248445 4948 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 00:07:34 crc kubenswrapper[4948]: I0312 00:07:34.246957 4948 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 00:07:35 crc kubenswrapper[4948]: I0312 00:07:35.248674 4948 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 00:07:35 crc kubenswrapper[4948]: I0312 00:07:35.312640 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:07:35 crc kubenswrapper[4948]: I0312 00:07:35.314860 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:07:35 crc kubenswrapper[4948]: I0312 00:07:35.314909 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:07:35 crc kubenswrapper[4948]: I0312 00:07:35.314932 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:07:35 crc kubenswrapper[4948]: I0312 00:07:35.315963 4948 scope.go:117] "RemoveContainer" containerID="e89f42aa5b3aa0882190e7501b56ba6f208e6c796e0b4041be478564c71b1c87" Mar 12 00:07:35 crc kubenswrapper[4948]: E0312 00:07:35.316349 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 12 00:07:36 crc kubenswrapper[4948]: I0312 00:07:36.249110 4948 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 00:07:37 crc kubenswrapper[4948]: I0312 00:07:37.201940 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:07:37 crc kubenswrapper[4948]: I0312 00:07:37.203527 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:07:37 crc kubenswrapper[4948]: I0312 00:07:37.203596 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:07:37 crc kubenswrapper[4948]: I0312 00:07:37.203617 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:07:37 crc kubenswrapper[4948]: I0312 00:07:37.203654 4948 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 12 00:07:37 crc kubenswrapper[4948]: E0312 00:07:37.210186 4948 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 12 00:07:37 crc kubenswrapper[4948]: E0312 00:07:37.211124 4948 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 12 00:07:37 crc kubenswrapper[4948]: I0312 00:07:37.248511 4948 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 00:07:38 crc kubenswrapper[4948]: I0312 00:07:38.248894 4948 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 00:07:39 crc kubenswrapper[4948]: I0312 00:07:39.248216 4948 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 00:07:40 crc kubenswrapper[4948]: I0312 00:07:40.246165 4948 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 00:07:41 crc kubenswrapper[4948]: I0312 00:07:41.249009 4948 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 00:07:41 crc kubenswrapper[4948]: E0312 00:07:41.413454 4948 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 12 00:07:42 crc kubenswrapper[4948]: I0312 00:07:42.246985 4948 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 00:07:43 crc kubenswrapper[4948]: I0312 00:07:43.247917 4948 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 00:07:44 crc kubenswrapper[4948]: I0312 00:07:44.211438 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:07:44 crc kubenswrapper[4948]: I0312 00:07:44.214121 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:07:44 crc kubenswrapper[4948]: I0312 00:07:44.214164 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:07:44 crc kubenswrapper[4948]: I0312 00:07:44.214174 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:07:44 crc kubenswrapper[4948]: I0312 00:07:44.214222 4948 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 12 00:07:44 crc kubenswrapper[4948]: E0312 00:07:44.224781 4948 kubelet_node_status.go:99] "Unable to register node with API server" err="nodes is forbidden: User \"system:anonymous\" cannot create resource \"nodes\" in API group \"\" at the cluster scope" node="crc" Mar 12 00:07:44 crc kubenswrapper[4948]: E0312 00:07:44.225329 4948 controller.go:145] "Failed to ensure lease exists, will retry" err="leases.coordination.k8s.io \"crc\" is forbidden: User \"system:anonymous\" cannot get resource \"leases\" in API group \"coordination.k8s.io\" in the namespace \"kube-node-lease\"" interval="7s" Mar 12 00:07:44 crc kubenswrapper[4948]: I0312 00:07:44.248620 4948 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 00:07:45 crc kubenswrapper[4948]: I0312 00:07:45.246025 4948 csi_plugin.go:884] Failed to contact API server when waiting for CSINode publishing: csinodes.storage.k8s.io "crc" is forbidden: User "system:anonymous" cannot get resource "csinodes" in API group "storage.k8s.io" at the cluster scope Mar 12 00:07:45 crc kubenswrapper[4948]: I0312 00:07:45.738944 4948 csr.go:261] certificate signing request csr-6rxks is approved, waiting to be issued Mar 12 00:07:45 crc kubenswrapper[4948]: I0312 00:07:45.755249 4948 csr.go:257] certificate signing request csr-6rxks is issued Mar 12 00:07:45 crc kubenswrapper[4948]: I0312 00:07:45.781468 4948 reconstruct.go:205] "DevicePaths of reconstructed volumes updated" Mar 12 00:07:46 crc kubenswrapper[4948]: I0312 00:07:46.077165 4948 transport.go:147] "Certificate rotation detected, shutting down client connections to start using new credentials" Mar 12 00:07:46 crc kubenswrapper[4948]: I0312 00:07:46.756603 4948 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-11-28 06:03:40.210650682 +0000 UTC Mar 12 00:07:46 crc kubenswrapper[4948]: I0312 00:07:46.756666 4948 certificate_manager.go:356] kubernetes.io/kube-apiserver-client-kubelet: Waiting 6269h55m53.453989962s for next certificate rotation Mar 12 00:07:49 crc kubenswrapper[4948]: I0312 00:07:49.312165 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:07:49 crc kubenswrapper[4948]: I0312 00:07:49.313878 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:07:49 crc kubenswrapper[4948]: I0312 00:07:49.314028 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:07:49 crc kubenswrapper[4948]: I0312 00:07:49.314114 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:07:49 crc kubenswrapper[4948]: I0312 00:07:49.314868 4948 scope.go:117] "RemoveContainer" containerID="e89f42aa5b3aa0882190e7501b56ba6f208e6c796e0b4041be478564c71b1c87" Mar 12 00:07:50 crc kubenswrapper[4948]: I0312 00:07:50.386878 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 12 00:07:50 crc kubenswrapper[4948]: I0312 00:07:50.387525 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/2.log" Mar 12 00:07:50 crc kubenswrapper[4948]: I0312 00:07:50.389097 4948 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="73834c8db23f6536034dce09e38b56665211238280bf9a4e1b13e74d81245a77" exitCode=255 Mar 12 00:07:50 crc kubenswrapper[4948]: I0312 00:07:50.389131 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerDied","Data":"73834c8db23f6536034dce09e38b56665211238280bf9a4e1b13e74d81245a77"} Mar 12 00:07:50 crc kubenswrapper[4948]: I0312 00:07:50.389161 4948 scope.go:117] "RemoveContainer" containerID="e89f42aa5b3aa0882190e7501b56ba6f208e6c796e0b4041be478564c71b1c87" Mar 12 00:07:50 crc kubenswrapper[4948]: I0312 00:07:50.389312 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:07:50 crc kubenswrapper[4948]: I0312 00:07:50.390081 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:07:50 crc kubenswrapper[4948]: I0312 00:07:50.390125 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:07:50 crc kubenswrapper[4948]: I0312 00:07:50.390143 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:07:50 crc kubenswrapper[4948]: I0312 00:07:50.391187 4948 scope.go:117] "RemoveContainer" containerID="73834c8db23f6536034dce09e38b56665211238280bf9a4e1b13e74d81245a77" Mar 12 00:07:50 crc kubenswrapper[4948]: E0312 00:07:50.391510 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 12 00:07:50 crc kubenswrapper[4948]: I0312 00:07:50.897658 4948 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 00:07:51 crc kubenswrapper[4948]: I0312 00:07:51.224953 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:07:51 crc kubenswrapper[4948]: I0312 00:07:51.226825 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:07:51 crc kubenswrapper[4948]: I0312 00:07:51.226859 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:07:51 crc kubenswrapper[4948]: I0312 00:07:51.226867 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:07:51 crc kubenswrapper[4948]: I0312 00:07:51.226952 4948 kubelet_node_status.go:76] "Attempting to register node" node="crc" Mar 12 00:07:51 crc kubenswrapper[4948]: I0312 00:07:51.235994 4948 kubelet_node_status.go:115] "Node was previously registered" node="crc" Mar 12 00:07:51 crc kubenswrapper[4948]: I0312 00:07:51.236143 4948 kubelet_node_status.go:79] "Successfully registered node" node="crc" Mar 12 00:07:51 crc kubenswrapper[4948]: E0312 00:07:51.236172 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Mar 12 00:07:51 crc kubenswrapper[4948]: I0312 00:07:51.239965 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:07:51 crc kubenswrapper[4948]: I0312 00:07:51.240006 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:07:51 crc kubenswrapper[4948]: I0312 00:07:51.240018 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:07:51 crc kubenswrapper[4948]: I0312 00:07:51.240035 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:07:51 crc kubenswrapper[4948]: I0312 00:07:51.240047 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:07:51Z","lastTransitionTime":"2026-03-12T00:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:07:51 crc kubenswrapper[4948]: E0312 00:07:51.264854 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:07:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:07:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:07:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:07:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17825588-29cf-4059-bab7-1b042b9bf1f3\\\",\\\"systemUUID\\\":\\\"b18e4bf2-d37e-4b80-8da3-cd3803754ba2\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 00:07:51 crc kubenswrapper[4948]: I0312 00:07:51.273432 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:07:51 crc kubenswrapper[4948]: I0312 00:07:51.273516 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:07:51 crc kubenswrapper[4948]: I0312 00:07:51.273533 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:07:51 crc kubenswrapper[4948]: I0312 00:07:51.273557 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:07:51 crc kubenswrapper[4948]: I0312 00:07:51.273575 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:07:51Z","lastTransitionTime":"2026-03-12T00:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:07:51 crc kubenswrapper[4948]: E0312 00:07:51.287913 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:07:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:07:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:07:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:07:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17825588-29cf-4059-bab7-1b042b9bf1f3\\\",\\\"systemUUID\\\":\\\"b18e4bf2-d37e-4b80-8da3-cd3803754ba2\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 00:07:51 crc kubenswrapper[4948]: I0312 00:07:51.296437 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:07:51 crc kubenswrapper[4948]: I0312 00:07:51.296500 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:07:51 crc kubenswrapper[4948]: I0312 00:07:51.296519 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:07:51 crc kubenswrapper[4948]: I0312 00:07:51.296544 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:07:51 crc kubenswrapper[4948]: I0312 00:07:51.296562 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:07:51Z","lastTransitionTime":"2026-03-12T00:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:07:51 crc kubenswrapper[4948]: E0312 00:07:51.310478 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:07:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:07:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:07:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:07:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17825588-29cf-4059-bab7-1b042b9bf1f3\\\",\\\"systemUUID\\\":\\\"b18e4bf2-d37e-4b80-8da3-cd3803754ba2\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 00:07:51 crc kubenswrapper[4948]: I0312 00:07:51.311659 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:07:51 crc kubenswrapper[4948]: I0312 00:07:51.312837 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:07:51 crc kubenswrapper[4948]: I0312 00:07:51.312876 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:07:51 crc kubenswrapper[4948]: I0312 00:07:51.312888 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:07:51 crc kubenswrapper[4948]: I0312 00:07:51.318042 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:07:51 crc kubenswrapper[4948]: I0312 00:07:51.318097 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:07:51 crc kubenswrapper[4948]: I0312 00:07:51.318115 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:07:51 crc kubenswrapper[4948]: I0312 00:07:51.318138 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:07:51 crc kubenswrapper[4948]: I0312 00:07:51.318154 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:07:51Z","lastTransitionTime":"2026-03-12T00:07:51Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:07:51 crc kubenswrapper[4948]: E0312 00:07:51.333114 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:07:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:07:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:51Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:07:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:51Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:07:51Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:51Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17825588-29cf-4059-bab7-1b042b9bf1f3\\\",\\\"systemUUID\\\":\\\"b18e4bf2-d37e-4b80-8da3-cd3803754ba2\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 00:07:51 crc kubenswrapper[4948]: E0312 00:07:51.333277 4948 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 12 00:07:51 crc kubenswrapper[4948]: E0312 00:07:51.333326 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:51 crc kubenswrapper[4948]: I0312 00:07:51.393394 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 12 00:07:51 crc kubenswrapper[4948]: I0312 00:07:51.395739 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:07:51 crc kubenswrapper[4948]: I0312 00:07:51.397049 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:07:51 crc kubenswrapper[4948]: I0312 00:07:51.397095 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:07:51 crc kubenswrapper[4948]: I0312 00:07:51.397110 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:07:51 crc kubenswrapper[4948]: I0312 00:07:51.397923 4948 scope.go:117] "RemoveContainer" containerID="73834c8db23f6536034dce09e38b56665211238280bf9a4e1b13e74d81245a77" Mar 12 00:07:51 crc kubenswrapper[4948]: E0312 00:07:51.398147 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 12 00:07:51 crc kubenswrapper[4948]: E0312 00:07:51.414889 4948 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 12 00:07:51 crc kubenswrapper[4948]: E0312 00:07:51.433777 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:51 crc kubenswrapper[4948]: E0312 00:07:51.534812 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:51 crc kubenswrapper[4948]: E0312 00:07:51.635555 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:51 crc kubenswrapper[4948]: E0312 00:07:51.736704 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:51 crc kubenswrapper[4948]: E0312 00:07:51.837468 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:51 crc kubenswrapper[4948]: E0312 00:07:51.938418 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:52 crc kubenswrapper[4948]: E0312 00:07:52.038686 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:52 crc kubenswrapper[4948]: E0312 00:07:52.139055 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:52 crc kubenswrapper[4948]: E0312 00:07:52.239134 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:52 crc kubenswrapper[4948]: E0312 00:07:52.339724 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:52 crc kubenswrapper[4948]: E0312 00:07:52.440439 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:52 crc kubenswrapper[4948]: E0312 00:07:52.540503 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:52 crc kubenswrapper[4948]: E0312 00:07:52.640766 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:52 crc kubenswrapper[4948]: E0312 00:07:52.741672 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:52 crc kubenswrapper[4948]: E0312 00:07:52.842077 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:52 crc kubenswrapper[4948]: E0312 00:07:52.942990 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:53 crc kubenswrapper[4948]: E0312 00:07:53.043932 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:53 crc kubenswrapper[4948]: E0312 00:07:53.145142 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:53 crc kubenswrapper[4948]: E0312 00:07:53.245736 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:53 crc kubenswrapper[4948]: E0312 00:07:53.346838 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:53 crc kubenswrapper[4948]: E0312 00:07:53.447993 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:53 crc kubenswrapper[4948]: E0312 00:07:53.548988 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:53 crc kubenswrapper[4948]: E0312 00:07:53.650046 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:53 crc kubenswrapper[4948]: E0312 00:07:53.750734 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:53 crc kubenswrapper[4948]: E0312 00:07:53.851889 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:53 crc kubenswrapper[4948]: E0312 00:07:53.952141 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:53 crc kubenswrapper[4948]: I0312 00:07:53.974384 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 00:07:53 crc kubenswrapper[4948]: I0312 00:07:53.974635 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:07:53 crc kubenswrapper[4948]: I0312 00:07:53.976343 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:07:53 crc kubenswrapper[4948]: I0312 00:07:53.976418 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:07:53 crc kubenswrapper[4948]: I0312 00:07:53.976441 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:07:53 crc kubenswrapper[4948]: I0312 00:07:53.977417 4948 scope.go:117] "RemoveContainer" containerID="73834c8db23f6536034dce09e38b56665211238280bf9a4e1b13e74d81245a77" Mar 12 00:07:53 crc kubenswrapper[4948]: E0312 00:07:53.977874 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 12 00:07:54 crc kubenswrapper[4948]: E0312 00:07:54.052781 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:54 crc kubenswrapper[4948]: E0312 00:07:54.152932 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:54 crc kubenswrapper[4948]: E0312 00:07:54.253720 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:54 crc kubenswrapper[4948]: E0312 00:07:54.354603 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:54 crc kubenswrapper[4948]: E0312 00:07:54.454982 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:54 crc kubenswrapper[4948]: E0312 00:07:54.555281 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:54 crc kubenswrapper[4948]: E0312 00:07:54.655753 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:54 crc kubenswrapper[4948]: E0312 00:07:54.756604 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:54 crc kubenswrapper[4948]: E0312 00:07:54.857573 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:54 crc kubenswrapper[4948]: E0312 00:07:54.958057 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:55 crc kubenswrapper[4948]: E0312 00:07:55.059093 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:55 crc kubenswrapper[4948]: E0312 00:07:55.159516 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:55 crc kubenswrapper[4948]: E0312 00:07:55.260383 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:55 crc kubenswrapper[4948]: E0312 00:07:55.360560 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:55 crc kubenswrapper[4948]: E0312 00:07:55.461628 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:55 crc kubenswrapper[4948]: E0312 00:07:55.562720 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:55 crc kubenswrapper[4948]: E0312 00:07:55.662985 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:55 crc kubenswrapper[4948]: E0312 00:07:55.763865 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:55 crc kubenswrapper[4948]: E0312 00:07:55.864920 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:55 crc kubenswrapper[4948]: E0312 00:07:55.965917 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:56 crc kubenswrapper[4948]: E0312 00:07:56.066563 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:56 crc kubenswrapper[4948]: E0312 00:07:56.167167 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:56 crc kubenswrapper[4948]: E0312 00:07:56.267709 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:56 crc kubenswrapper[4948]: E0312 00:07:56.368539 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:56 crc kubenswrapper[4948]: E0312 00:07:56.468627 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:56 crc kubenswrapper[4948]: E0312 00:07:56.568863 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:56 crc kubenswrapper[4948]: E0312 00:07:56.669409 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:56 crc kubenswrapper[4948]: E0312 00:07:56.770155 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:56 crc kubenswrapper[4948]: E0312 00:07:56.871194 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:56 crc kubenswrapper[4948]: E0312 00:07:56.972345 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:57 crc kubenswrapper[4948]: E0312 00:07:57.073342 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:57 crc kubenswrapper[4948]: E0312 00:07:57.173815 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:57 crc kubenswrapper[4948]: E0312 00:07:57.274250 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:57 crc kubenswrapper[4948]: E0312 00:07:57.374773 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:57 crc kubenswrapper[4948]: E0312 00:07:57.475264 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:57 crc kubenswrapper[4948]: E0312 00:07:57.576422 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:57 crc kubenswrapper[4948]: E0312 00:07:57.677189 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:57 crc kubenswrapper[4948]: E0312 00:07:57.778334 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:57 crc kubenswrapper[4948]: E0312 00:07:57.879206 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:57 crc kubenswrapper[4948]: E0312 00:07:57.980084 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:58 crc kubenswrapper[4948]: E0312 00:07:58.080963 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:58 crc kubenswrapper[4948]: E0312 00:07:58.181409 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:58 crc kubenswrapper[4948]: E0312 00:07:58.282088 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:58 crc kubenswrapper[4948]: E0312 00:07:58.382889 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:58 crc kubenswrapper[4948]: E0312 00:07:58.483956 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:58 crc kubenswrapper[4948]: E0312 00:07:58.584934 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:58 crc kubenswrapper[4948]: E0312 00:07:58.685433 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:58 crc kubenswrapper[4948]: E0312 00:07:58.785610 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:58 crc kubenswrapper[4948]: E0312 00:07:58.886663 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:58 crc kubenswrapper[4948]: E0312 00:07:58.987655 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:59 crc kubenswrapper[4948]: E0312 00:07:59.088357 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:59 crc kubenswrapper[4948]: E0312 00:07:59.189122 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:59 crc kubenswrapper[4948]: E0312 00:07:59.290024 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:59 crc kubenswrapper[4948]: E0312 00:07:59.390820 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:59 crc kubenswrapper[4948]: E0312 00:07:59.491284 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:59 crc kubenswrapper[4948]: E0312 00:07:59.592397 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:59 crc kubenswrapper[4948]: E0312 00:07:59.692841 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:59 crc kubenswrapper[4948]: E0312 00:07:59.793938 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:59 crc kubenswrapper[4948]: E0312 00:07:59.894909 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:07:59 crc kubenswrapper[4948]: E0312 00:07:59.996045 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:00 crc kubenswrapper[4948]: E0312 00:08:00.096583 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:00 crc kubenswrapper[4948]: E0312 00:08:00.197651 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:00 crc kubenswrapper[4948]: E0312 00:08:00.297997 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:00 crc kubenswrapper[4948]: E0312 00:08:00.398496 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:00 crc kubenswrapper[4948]: E0312 00:08:00.498665 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:00 crc kubenswrapper[4948]: E0312 00:08:00.599679 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:00 crc kubenswrapper[4948]: E0312 00:08:00.700000 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:00 crc kubenswrapper[4948]: E0312 00:08:00.800448 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:00 crc kubenswrapper[4948]: E0312 00:08:00.900600 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:01 crc kubenswrapper[4948]: E0312 00:08:01.001257 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:01 crc kubenswrapper[4948]: E0312 00:08:01.101694 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:01 crc kubenswrapper[4948]: E0312 00:08:01.202023 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:01 crc kubenswrapper[4948]: E0312 00:08:01.302774 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:01 crc kubenswrapper[4948]: E0312 00:08:01.403440 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:01 crc kubenswrapper[4948]: E0312 00:08:01.407608 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Mar 12 00:08:01 crc kubenswrapper[4948]: E0312 00:08:01.415612 4948 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 12 00:08:01 crc kubenswrapper[4948]: I0312 00:08:01.416213 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:01 crc kubenswrapper[4948]: I0312 00:08:01.416267 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:01 crc kubenswrapper[4948]: I0312 00:08:01.416287 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:01 crc kubenswrapper[4948]: I0312 00:08:01.416341 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:01 crc kubenswrapper[4948]: I0312 00:08:01.416359 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:01Z","lastTransitionTime":"2026-03-12T00:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:01 crc kubenswrapper[4948]: E0312 00:08:01.431126 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17825588-29cf-4059-bab7-1b042b9bf1f3\\\",\\\"systemUUID\\\":\\\"b18e4bf2-d37e-4b80-8da3-cd3803754ba2\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 00:08:01 crc kubenswrapper[4948]: I0312 00:08:01.438835 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:01 crc kubenswrapper[4948]: I0312 00:08:01.438881 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:01 crc kubenswrapper[4948]: I0312 00:08:01.438898 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:01 crc kubenswrapper[4948]: I0312 00:08:01.438921 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:01 crc kubenswrapper[4948]: I0312 00:08:01.438938 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:01Z","lastTransitionTime":"2026-03-12T00:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:01 crc kubenswrapper[4948]: E0312 00:08:01.454237 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17825588-29cf-4059-bab7-1b042b9bf1f3\\\",\\\"systemUUID\\\":\\\"b18e4bf2-d37e-4b80-8da3-cd3803754ba2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 00:08:01 crc kubenswrapper[4948]: I0312 00:08:01.463710 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:01 crc kubenswrapper[4948]: I0312 00:08:01.463899 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:01 crc kubenswrapper[4948]: I0312 00:08:01.463933 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:01 crc kubenswrapper[4948]: I0312 00:08:01.464453 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:01 crc kubenswrapper[4948]: I0312 00:08:01.464609 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:01Z","lastTransitionTime":"2026-03-12T00:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:01 crc kubenswrapper[4948]: E0312 00:08:01.477057 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17825588-29cf-4059-bab7-1b042b9bf1f3\\\",\\\"systemUUID\\\":\\\"b18e4bf2-d37e-4b80-8da3-cd3803754ba2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 00:08:01 crc kubenswrapper[4948]: I0312 00:08:01.486453 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:01 crc kubenswrapper[4948]: I0312 00:08:01.486488 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:01 crc kubenswrapper[4948]: I0312 00:08:01.486515 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:01 crc kubenswrapper[4948]: I0312 00:08:01.486531 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:01 crc kubenswrapper[4948]: I0312 00:08:01.486540 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:01Z","lastTransitionTime":"2026-03-12T00:08:01Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:01 crc kubenswrapper[4948]: E0312 00:08:01.499731 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:01Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:01Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:01Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:01Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17825588-29cf-4059-bab7-1b042b9bf1f3\\\",\\\"systemUUID\\\":\\\"b18e4bf2-d37e-4b80-8da3-cd3803754ba2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 00:08:01 crc kubenswrapper[4948]: E0312 00:08:01.499862 4948 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 12 00:08:01 crc kubenswrapper[4948]: E0312 00:08:01.503809 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:01 crc kubenswrapper[4948]: E0312 00:08:01.604382 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:01 crc kubenswrapper[4948]: E0312 00:08:01.704851 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:01 crc kubenswrapper[4948]: I0312 00:08:01.743062 4948 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Mar 12 00:08:01 crc kubenswrapper[4948]: E0312 00:08:01.805991 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:01 crc kubenswrapper[4948]: E0312 00:08:01.906719 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:02 crc kubenswrapper[4948]: E0312 00:08:02.007881 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:02 crc kubenswrapper[4948]: E0312 00:08:02.108960 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:02 crc kubenswrapper[4948]: E0312 00:08:02.209393 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:02 crc kubenswrapper[4948]: E0312 00:08:02.309875 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:02 crc kubenswrapper[4948]: E0312 00:08:02.411047 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:02 crc kubenswrapper[4948]: E0312 00:08:02.511419 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:02 crc kubenswrapper[4948]: E0312 00:08:02.611769 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:02 crc kubenswrapper[4948]: E0312 00:08:02.712619 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:02 crc kubenswrapper[4948]: E0312 00:08:02.812976 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:02 crc kubenswrapper[4948]: E0312 00:08:02.913398 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:03 crc kubenswrapper[4948]: E0312 00:08:03.013560 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:03 crc kubenswrapper[4948]: E0312 00:08:03.113894 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:03 crc kubenswrapper[4948]: E0312 00:08:03.214932 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:03 crc kubenswrapper[4948]: I0312 00:08:03.312009 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:08:03 crc kubenswrapper[4948]: I0312 00:08:03.313391 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:03 crc kubenswrapper[4948]: I0312 00:08:03.313486 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:03 crc kubenswrapper[4948]: I0312 00:08:03.313509 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:03 crc kubenswrapper[4948]: E0312 00:08:03.315868 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:03 crc kubenswrapper[4948]: E0312 00:08:03.416956 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:03 crc kubenswrapper[4948]: E0312 00:08:03.517421 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:03 crc kubenswrapper[4948]: E0312 00:08:03.617541 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:03 crc kubenswrapper[4948]: E0312 00:08:03.718534 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:03 crc kubenswrapper[4948]: E0312 00:08:03.819929 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:03 crc kubenswrapper[4948]: E0312 00:08:03.920885 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:04 crc kubenswrapper[4948]: E0312 00:08:04.021768 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:04 crc kubenswrapper[4948]: E0312 00:08:04.123041 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:04 crc kubenswrapper[4948]: E0312 00:08:04.223851 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:04 crc kubenswrapper[4948]: E0312 00:08:04.324934 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:04 crc kubenswrapper[4948]: E0312 00:08:04.425409 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:04 crc kubenswrapper[4948]: E0312 00:08:04.525537 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:04 crc kubenswrapper[4948]: E0312 00:08:04.626660 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:04 crc kubenswrapper[4948]: E0312 00:08:04.727252 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:04 crc kubenswrapper[4948]: E0312 00:08:04.828424 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:04 crc kubenswrapper[4948]: E0312 00:08:04.929385 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:05 crc kubenswrapper[4948]: E0312 00:08:05.030267 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:05 crc kubenswrapper[4948]: E0312 00:08:05.131485 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:05 crc kubenswrapper[4948]: E0312 00:08:05.231867 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:05 crc kubenswrapper[4948]: E0312 00:08:05.332820 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:05 crc kubenswrapper[4948]: E0312 00:08:05.433797 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:05 crc kubenswrapper[4948]: E0312 00:08:05.534823 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:05 crc kubenswrapper[4948]: E0312 00:08:05.635154 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:05 crc kubenswrapper[4948]: E0312 00:08:05.736127 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:05 crc kubenswrapper[4948]: E0312 00:08:05.837294 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:05 crc kubenswrapper[4948]: E0312 00:08:05.938539 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:06 crc kubenswrapper[4948]: E0312 00:08:06.039637 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:06 crc kubenswrapper[4948]: E0312 00:08:06.139784 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:06 crc kubenswrapper[4948]: E0312 00:08:06.239916 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:06 crc kubenswrapper[4948]: E0312 00:08:06.340764 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:06 crc kubenswrapper[4948]: E0312 00:08:06.441785 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:06 crc kubenswrapper[4948]: E0312 00:08:06.541984 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:06 crc kubenswrapper[4948]: E0312 00:08:06.642334 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:06 crc kubenswrapper[4948]: E0312 00:08:06.743162 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:06 crc kubenswrapper[4948]: E0312 00:08:06.843397 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:06 crc kubenswrapper[4948]: E0312 00:08:06.944514 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:07 crc kubenswrapper[4948]: E0312 00:08:07.045289 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:07 crc kubenswrapper[4948]: E0312 00:08:07.145580 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:07 crc kubenswrapper[4948]: E0312 00:08:07.245768 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:07 crc kubenswrapper[4948]: I0312 00:08:07.312017 4948 kubelet_node_status.go:401] "Setting node annotation to enable volume controller attach/detach" Mar 12 00:08:07 crc kubenswrapper[4948]: I0312 00:08:07.313425 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:07 crc kubenswrapper[4948]: I0312 00:08:07.313559 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:07 crc kubenswrapper[4948]: I0312 00:08:07.313833 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:07 crc kubenswrapper[4948]: I0312 00:08:07.314731 4948 scope.go:117] "RemoveContainer" containerID="73834c8db23f6536034dce09e38b56665211238280bf9a4e1b13e74d81245a77" Mar 12 00:08:07 crc kubenswrapper[4948]: E0312 00:08:07.315016 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 12 00:08:07 crc kubenswrapper[4948]: E0312 00:08:07.346276 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:07 crc kubenswrapper[4948]: E0312 00:08:07.447226 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:07 crc kubenswrapper[4948]: E0312 00:08:07.548136 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:07 crc kubenswrapper[4948]: E0312 00:08:07.648481 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:07 crc kubenswrapper[4948]: I0312 00:08:07.706578 4948 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Mar 12 00:08:07 crc kubenswrapper[4948]: E0312 00:08:07.749153 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:07 crc kubenswrapper[4948]: E0312 00:08:07.850247 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:07 crc kubenswrapper[4948]: E0312 00:08:07.951179 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:08 crc kubenswrapper[4948]: E0312 00:08:08.052259 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:08 crc kubenswrapper[4948]: E0312 00:08:08.152379 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:08 crc kubenswrapper[4948]: E0312 00:08:08.253092 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:08 crc kubenswrapper[4948]: E0312 00:08:08.354061 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:08 crc kubenswrapper[4948]: E0312 00:08:08.455148 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:08 crc kubenswrapper[4948]: E0312 00:08:08.556100 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:08 crc kubenswrapper[4948]: E0312 00:08:08.657162 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:08 crc kubenswrapper[4948]: E0312 00:08:08.758083 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:08 crc kubenswrapper[4948]: E0312 00:08:08.858470 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:08 crc kubenswrapper[4948]: E0312 00:08:08.959564 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:09 crc kubenswrapper[4948]: E0312 00:08:09.060480 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:09 crc kubenswrapper[4948]: E0312 00:08:09.161029 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:09 crc kubenswrapper[4948]: E0312 00:08:09.262174 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:09 crc kubenswrapper[4948]: E0312 00:08:09.362765 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:09 crc kubenswrapper[4948]: E0312 00:08:09.463489 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:09 crc kubenswrapper[4948]: E0312 00:08:09.564363 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:09 crc kubenswrapper[4948]: E0312 00:08:09.664939 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:09 crc kubenswrapper[4948]: E0312 00:08:09.766048 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:09 crc kubenswrapper[4948]: E0312 00:08:09.867069 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:09 crc kubenswrapper[4948]: E0312 00:08:09.968132 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:10 crc kubenswrapper[4948]: E0312 00:08:10.068946 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:10 crc kubenswrapper[4948]: E0312 00:08:10.169490 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:10 crc kubenswrapper[4948]: E0312 00:08:10.270163 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:10 crc kubenswrapper[4948]: E0312 00:08:10.370612 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:10 crc kubenswrapper[4948]: E0312 00:08:10.470915 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:10 crc kubenswrapper[4948]: E0312 00:08:10.571563 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:10 crc kubenswrapper[4948]: E0312 00:08:10.672042 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:10 crc kubenswrapper[4948]: I0312 00:08:10.676785 4948 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Mar 12 00:08:10 crc kubenswrapper[4948]: E0312 00:08:10.772421 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:10 crc kubenswrapper[4948]: E0312 00:08:10.873359 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:10 crc kubenswrapper[4948]: E0312 00:08:10.973945 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:11 crc kubenswrapper[4948]: E0312 00:08:11.074487 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:11 crc kubenswrapper[4948]: E0312 00:08:11.175736 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:11 crc kubenswrapper[4948]: E0312 00:08:11.276417 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:11 crc kubenswrapper[4948]: E0312 00:08:11.376744 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:11 crc kubenswrapper[4948]: E0312 00:08:11.416148 4948 eviction_manager.go:285] "Eviction manager: failed to get summary stats" err="failed to get node info: node \"crc\" not found" Mar 12 00:08:11 crc kubenswrapper[4948]: E0312 00:08:11.477795 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:11 crc kubenswrapper[4948]: E0312 00:08:11.579004 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:11 crc kubenswrapper[4948]: E0312 00:08:11.593363 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="error getting node \"crc\": node \"crc\" not found" Mar 12 00:08:11 crc kubenswrapper[4948]: I0312 00:08:11.599003 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:11 crc kubenswrapper[4948]: I0312 00:08:11.599069 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:11 crc kubenswrapper[4948]: I0312 00:08:11.599093 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:11 crc kubenswrapper[4948]: I0312 00:08:11.599124 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:11 crc kubenswrapper[4948]: I0312 00:08:11.599146 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:11Z","lastTransitionTime":"2026-03-12T00:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:11 crc kubenswrapper[4948]: E0312 00:08:11.615110 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17825588-29cf-4059-bab7-1b042b9bf1f3\\\",\\\"systemUUID\\\":\\\"b18e4bf2-d37e-4b80-8da3-cd3803754ba2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 00:08:11 crc kubenswrapper[4948]: I0312 00:08:11.620151 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:11 crc kubenswrapper[4948]: I0312 00:08:11.620402 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:11 crc kubenswrapper[4948]: I0312 00:08:11.620618 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:11 crc kubenswrapper[4948]: I0312 00:08:11.620835 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:11 crc kubenswrapper[4948]: I0312 00:08:11.620990 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:11Z","lastTransitionTime":"2026-03-12T00:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:11 crc kubenswrapper[4948]: E0312 00:08:11.637059 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17825588-29cf-4059-bab7-1b042b9bf1f3\\\",\\\"systemUUID\\\":\\\"b18e4bf2-d37e-4b80-8da3-cd3803754ba2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 00:08:11 crc kubenswrapper[4948]: I0312 00:08:11.641949 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:11 crc kubenswrapper[4948]: I0312 00:08:11.641985 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:11 crc kubenswrapper[4948]: I0312 00:08:11.642000 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:11 crc kubenswrapper[4948]: I0312 00:08:11.642023 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:11 crc kubenswrapper[4948]: I0312 00:08:11.642040 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:11Z","lastTransitionTime":"2026-03-12T00:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:11 crc kubenswrapper[4948]: E0312 00:08:11.658127 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17825588-29cf-4059-bab7-1b042b9bf1f3\\\",\\\"systemUUID\\\":\\\"b18e4bf2-d37e-4b80-8da3-cd3803754ba2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 00:08:11 crc kubenswrapper[4948]: I0312 00:08:11.662984 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:11 crc kubenswrapper[4948]: I0312 00:08:11.663196 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:11 crc kubenswrapper[4948]: I0312 00:08:11.663379 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:11 crc kubenswrapper[4948]: I0312 00:08:11.663528 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:11 crc kubenswrapper[4948]: I0312 00:08:11.663664 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:11Z","lastTransitionTime":"2026-03-12T00:08:11Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:11 crc kubenswrapper[4948]: E0312 00:08:11.679180 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:11Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:11Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:11Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:11Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17825588-29cf-4059-bab7-1b042b9bf1f3\\\",\\\"systemUUID\\\":\\\"b18e4bf2-d37e-4b80-8da3-cd3803754ba2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 00:08:11 crc kubenswrapper[4948]: E0312 00:08:11.679438 4948 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 12 00:08:11 crc kubenswrapper[4948]: E0312 00:08:11.679476 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:11 crc kubenswrapper[4948]: E0312 00:08:11.779613 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:11 crc kubenswrapper[4948]: E0312 00:08:11.880738 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:11 crc kubenswrapper[4948]: E0312 00:08:11.981637 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:12 crc kubenswrapper[4948]: E0312 00:08:12.082600 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:12 crc kubenswrapper[4948]: E0312 00:08:12.183604 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:12 crc kubenswrapper[4948]: E0312 00:08:12.284011 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:12 crc kubenswrapper[4948]: E0312 00:08:12.385174 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:12 crc kubenswrapper[4948]: E0312 00:08:12.486530 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:12 crc kubenswrapper[4948]: E0312 00:08:12.587838 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:12 crc kubenswrapper[4948]: E0312 00:08:12.688878 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:12 crc kubenswrapper[4948]: E0312 00:08:12.789628 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:12 crc kubenswrapper[4948]: E0312 00:08:12.889802 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:12 crc kubenswrapper[4948]: E0312 00:08:12.989951 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:13 crc kubenswrapper[4948]: E0312 00:08:13.090738 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:13 crc kubenswrapper[4948]: E0312 00:08:13.191090 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:13 crc kubenswrapper[4948]: E0312 00:08:13.292109 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:13 crc kubenswrapper[4948]: E0312 00:08:13.393205 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:13 crc kubenswrapper[4948]: E0312 00:08:13.493483 4948 kubelet_node_status.go:503] "Error getting the current node from lister" err="node \"crc\" not found" Mar 12 00:08:13 crc kubenswrapper[4948]: I0312 00:08:13.564889 4948 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Mar 12 00:08:13 crc kubenswrapper[4948]: I0312 00:08:13.595743 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:13 crc kubenswrapper[4948]: I0312 00:08:13.595817 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:13 crc kubenswrapper[4948]: I0312 00:08:13.595841 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:13 crc kubenswrapper[4948]: I0312 00:08:13.595871 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:13 crc kubenswrapper[4948]: I0312 00:08:13.595895 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:13Z","lastTransitionTime":"2026-03-12T00:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:13 crc kubenswrapper[4948]: I0312 00:08:13.698861 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:13 crc kubenswrapper[4948]: I0312 00:08:13.698928 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:13 crc kubenswrapper[4948]: I0312 00:08:13.698950 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:13 crc kubenswrapper[4948]: I0312 00:08:13.698978 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:13 crc kubenswrapper[4948]: I0312 00:08:13.699003 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:13Z","lastTransitionTime":"2026-03-12T00:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:13 crc kubenswrapper[4948]: I0312 00:08:13.801686 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:13 crc kubenswrapper[4948]: I0312 00:08:13.801949 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:13 crc kubenswrapper[4948]: I0312 00:08:13.802046 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:13 crc kubenswrapper[4948]: I0312 00:08:13.802130 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:13 crc kubenswrapper[4948]: I0312 00:08:13.802187 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:13Z","lastTransitionTime":"2026-03-12T00:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:13 crc kubenswrapper[4948]: I0312 00:08:13.904441 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:13 crc kubenswrapper[4948]: I0312 00:08:13.904488 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:13 crc kubenswrapper[4948]: I0312 00:08:13.904499 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:13 crc kubenswrapper[4948]: I0312 00:08:13.904514 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:13 crc kubenswrapper[4948]: I0312 00:08:13.904530 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:13Z","lastTransitionTime":"2026-03-12T00:08:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.007406 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.007480 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.007504 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.007534 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.007611 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:14Z","lastTransitionTime":"2026-03-12T00:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.110235 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.110296 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.110370 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.110429 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.110455 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:14Z","lastTransitionTime":"2026-03-12T00:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.213808 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.213879 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.213902 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.213932 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.213954 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:14Z","lastTransitionTime":"2026-03-12T00:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.293160 4948 apiserver.go:52] "Watching apiserver" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.300556 4948 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.301746 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/network-metrics-daemon-mqmtd","openshift-network-console/networking-console-plugin-85b44fc459-gdk6g","openshift-network-diagnostics/network-check-target-xd92c","openshift-image-registry/node-ca-njc9c","openshift-machine-config-operator/machine-config-daemon-m4xwc","openshift-multus/multus-additional-cni-plugins-mfp2k","openshift-multus/multus-bwpqp","openshift-network-operator/network-operator-58b4c7f79c-55gtf","openshift-dns/node-resolver-9s5fr","openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xdd8","openshift-ovn-kubernetes/ovnkube-node-f7mp5","openshift-network-diagnostics/network-check-source-55646444c4-trplf","openshift-network-node-identity/network-node-identity-vrzqb","openshift-network-operator/iptables-alerter-4ln5h"] Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.302330 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.302366 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:08:14 crc kubenswrapper[4948]: E0312 00:08:14.302419 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 00:08:14 crc kubenswrapper[4948]: E0312 00:08:14.302486 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.302821 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.303056 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.303078 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:08:14 crc kubenswrapper[4948]: E0312 00:08:14.303195 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.303498 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.303577 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-bwpqp" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.303785 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.303936 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-9s5fr" Mar 12 00:08:14 crc kubenswrapper[4948]: E0312 00:08:14.304031 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mqmtd" podUID="be985ff1-de8a-431f-a230-67894ebe52d5" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.304067 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xdd8" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.304258 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.304868 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-mfp2k" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.304951 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-njc9c" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.304980 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.309630 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.309666 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.309699 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.309908 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.309942 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.309951 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.310025 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.310034 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.309637 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.310971 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.311079 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.311108 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.311273 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.311550 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.311636 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.311648 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.311689 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.311870 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.311902 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.311934 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.312035 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.312067 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.312090 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.312135 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.312153 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.312204 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.312208 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.316914 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.316967 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.316979 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.317001 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.317018 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:14Z","lastTransitionTime":"2026-03-12T00:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.320162 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.323860 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.324115 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.325360 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.340160 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.340560 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.340961 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.341039 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.341142 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.351192 4948 desired_state_of_world_populator.go:154] "Finished populating initial desired state of world" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.358618 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.358777 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.358874 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.359054 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.359196 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.359286 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.359402 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.359499 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.359606 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.359694 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.359776 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.359858 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.359951 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.360037 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.360114 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.360196 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.360272 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.360385 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.360523 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.360647 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.360737 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.360860 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.360947 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.361031 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.361117 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.361210 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.361312 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.361418 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.361511 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.361607 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.361692 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.361775 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.361859 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.361945 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.362039 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") pod \"44663579-783b-4372-86d6-acf235a62d72\" (UID: \"44663579-783b-4372-86d6-acf235a62d72\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.362124 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") pod \"6731426b-95fe-49ff-bb5f-40441049fde2\" (UID: \"6731426b-95fe-49ff-bb5f-40441049fde2\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.362208 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.362294 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.362403 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.362492 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.362592 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.362690 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.362773 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") pod \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\" (UID: \"96b93a3a-6083-4aea-8eab-fe1aa8245ad9\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.362852 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.362938 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.363018 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.363102 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.363179 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.363264 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.363431 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.363516 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.363600 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.363687 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.363798 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.363892 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.363975 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.364063 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.364155 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.364241 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.364352 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.364451 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.364541 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.364652 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") pod \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\" (UID: \"b11524ee-3fca-4b1b-9cdf-6da289fdbc7d\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.364750 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.364846 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.364932 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.365140 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.365237 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") pod \"925f1c65-6136-48ba-85aa-3a3b50560753\" (UID: \"925f1c65-6136-48ba-85aa-3a3b50560753\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.367339 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca" (OuterVolumeSpecName: "image-import-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "image-import-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.367666 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh" (OuterVolumeSpecName: "kube-api-access-x7zkh") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "kube-api-access-x7zkh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.367962 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit" (OuterVolumeSpecName: "audit") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "audit". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.367982 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz" (OuterVolumeSpecName: "kube-api-access-6g6sz") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "kube-api-access-6g6sz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.368173 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh" (OuterVolumeSpecName: "kube-api-access-2w9zh") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "kube-api-access-2w9zh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.368361 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.368742 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.368955 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key" (OuterVolumeSpecName: "signing-key") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-key". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.368993 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.369163 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.369367 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh" (OuterVolumeSpecName: "kube-api-access-x4zgh") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "kube-api-access-x4zgh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.369537 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf" (OuterVolumeSpecName: "kube-api-access-v47cf") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "kube-api-access-v47cf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.369664 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config" (OuterVolumeSpecName: "config") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.369832 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.369973 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.370022 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4" (OuterVolumeSpecName: "kube-api-access-w4xd4") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "kube-api-access-w4xd4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.370215 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.370283 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7" (OuterVolumeSpecName: "kube-api-access-sb6h7") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "kube-api-access-sb6h7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.370405 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.370646 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.370823 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.370895 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc" (OuterVolumeSpecName: "kube-api-access-vt5rc") pod "44663579-783b-4372-86d6-acf235a62d72" (UID: "44663579-783b-4372-86d6-acf235a62d72"). InnerVolumeSpecName "kube-api-access-vt5rc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.371085 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7" (OuterVolumeSpecName: "kube-api-access-nzwt7") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "kube-api-access-nzwt7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.371105 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls" (OuterVolumeSpecName: "control-plane-machine-set-operator-tls") pod "6731426b-95fe-49ff-bb5f-40441049fde2" (UID: "6731426b-95fe-49ff-bb5f-40441049fde2"). InnerVolumeSpecName "control-plane-machine-set-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.371143 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config" (OuterVolumeSpecName: "config") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.371228 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.371357 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp" (OuterVolumeSpecName: "kube-api-access-qs4fp") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "kube-api-access-qs4fp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.371435 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv" (OuterVolumeSpecName: "kube-api-access-zkvpv") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "kube-api-access-zkvpv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.371485 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.371703 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.371908 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn" (OuterVolumeSpecName: "kube-api-access-lz9wn") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "kube-api-access-lz9wn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.371938 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert" (OuterVolumeSpecName: "ovn-control-plane-metrics-cert") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovn-control-plane-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.371988 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities" (OuterVolumeSpecName: "utilities") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.372146 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.372199 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl" (OuterVolumeSpecName: "kube-api-access-xcphl") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "kube-api-access-xcphl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.372236 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.372365 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs" (OuterVolumeSpecName: "metrics-certs") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "metrics-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.372419 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct" (OuterVolumeSpecName: "kube-api-access-cfbct") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "kube-api-access-cfbct". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.372428 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m" (OuterVolumeSpecName: "kube-api-access-gf66m") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "kube-api-access-gf66m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.372439 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.372568 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.372793 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.372963 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.373118 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.373262 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config" (OuterVolumeSpecName: "config") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: E0312 00:08:14.373313 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 00:08:14.873245504 +0000 UTC m=+94.328849252 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.384080 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.385064 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.386777 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.388909 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.389128 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.389241 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.389726 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.390135 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.393800 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.400438 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") pod \"49ef4625-1d3a-4a9f-b595-c2433d32326d\" (UID: \"49ef4625-1d3a-4a9f-b595-c2433d32326d\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.400843 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.400943 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") pod \"e7e6199b-1264-4501-8953-767f51328d08\" (UID: \"e7e6199b-1264-4501-8953-767f51328d08\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.401132 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.401389 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.401474 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") pod \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\" (UID: \"210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.401576 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.401649 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.401927 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.373335 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle" (OuterVolumeSpecName: "signing-cabundle") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "signing-cabundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.373347 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config" (OuterVolumeSpecName: "console-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.373490 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "96b93a3a-6083-4aea-8eab-fe1aa8245ad9" (UID: "96b93a3a-6083-4aea-8eab-fe1aa8245ad9"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.373532 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls" (OuterVolumeSpecName: "image-registry-operator-tls") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "image-registry-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.373726 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config" (OuterVolumeSpecName: "config") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.373827 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.373927 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj" (OuterVolumeSpecName: "kube-api-access-4d4hj") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "kube-api-access-4d4hj". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.374149 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.374211 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca" (OuterVolumeSpecName: "client-ca") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.374276 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config" (OuterVolumeSpecName: "config") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.374500 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.374900 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities" (OuterVolumeSpecName: "utilities") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.374961 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config" (OuterVolumeSpecName: "config") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.375261 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.375336 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88" (OuterVolumeSpecName: "kube-api-access-lzf88") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "kube-api-access-lzf88". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.375423 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.376232 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52" (OuterVolumeSpecName: "kube-api-access-s4n52") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "kube-api-access-s4n52". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.377005 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "925f1c65-6136-48ba-85aa-3a3b50560753" (UID: "925f1c65-6136-48ba-85aa-3a3b50560753"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.377125 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config" (OuterVolumeSpecName: "config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.376680 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config" (OuterVolumeSpecName: "config") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.381604 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca" (OuterVolumeSpecName: "serviceca") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.384999 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert" (OuterVolumeSpecName: "console-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.386687 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config" (OuterVolumeSpecName: "config") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.389680 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.390659 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.390730 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.393759 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.394137 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate" (OuterVolumeSpecName: "default-certificate") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "default-certificate". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.394168 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert" (OuterVolumeSpecName: "srv-cert") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "srv-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.400703 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v" (OuterVolumeSpecName: "kube-api-access-pjr6v") pod "49ef4625-1d3a-4a9f-b595-c2433d32326d" (UID: "49ef4625-1d3a-4a9f-b595-c2433d32326d"). InnerVolumeSpecName "kube-api-access-pjr6v". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.401138 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token" (OuterVolumeSpecName: "node-bootstrap-token") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "node-bootstrap-token". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.401475 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config" (OuterVolumeSpecName: "console-oauth-config") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "console-oauth-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.401812 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config" (OuterVolumeSpecName: "config") pod "e7e6199b-1264-4501-8953-767f51328d08" (UID: "e7e6199b-1264-4501-8953-767f51328d08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.401857 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" (UID: "210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.387831 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.403421 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.403489 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.403507 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") pod \"496e6271-fb68-4057-954e-a0d97a4afa3f\" (UID: \"496e6271-fb68-4057-954e-a0d97a4afa3f\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.403524 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.403541 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.403558 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.403575 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.403594 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.403609 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") pod \"5b88f790-22fa-440e-b583-365168c0b23d\" (UID: \"5b88f790-22fa-440e-b583-365168c0b23d\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.403628 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.403644 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.403659 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") pod \"4bb40260-dbaa-4fb0-84df-5e680505d512\" (UID: \"4bb40260-dbaa-4fb0-84df-5e680505d512\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.403673 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.403689 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.403705 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") pod \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\" (UID: \"c03ee662-fb2f-4fc4-a2c1-af487c19d254\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.403722 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.403738 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") pod \"a31745f5-9847-4afe-82a5-3161cc66ca93\" (UID: \"a31745f5-9847-4afe-82a5-3161cc66ca93\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.403754 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") pod \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\" (UID: \"3cb93b32-e0ae-4377-b9c8-fdb9842c6d59\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.403771 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.403790 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.403805 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.403822 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") pod \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\" (UID: \"8cea82b4-6893-4ddc-af9f-1bb5ae425c5b\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.403837 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.403853 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.403870 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.403886 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.403907 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.403932 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.403948 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.403964 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.403981 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.403998 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") pod \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\" (UID: \"f88749ec-7931-4ee7-b3fc-1ec5e11f92e9\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.404017 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.404034 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") pod \"6402fda4-df10-493c-b4e5-d0569419652d\" (UID: \"6402fda4-df10-493c-b4e5-d0569419652d\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.404052 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.404069 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.404087 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") pod \"5fe579f8-e8a6-4643-bce5-a661393c4dde\" (UID: \"5fe579f8-e8a6-4643-bce5-a661393c4dde\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.404105 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.404122 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") pod \"bf126b07-da06-4140-9a57-dfd54fc6b486\" (UID: \"bf126b07-da06-4140-9a57-dfd54fc6b486\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.404140 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") pod \"efdd0498-1daa-4136-9a4a-3b948c2293fc\" (UID: \"efdd0498-1daa-4136-9a4a-3b948c2293fc\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.404156 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.404175 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") pod \"5225d0e4-402f-4861-b410-819f433b1803\" (UID: \"5225d0e4-402f-4861-b410-819f433b1803\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.404192 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") pod \"0b78653f-4ff9-4508-8672-245ed9b561e3\" (UID: \"0b78653f-4ff9-4508-8672-245ed9b561e3\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.404208 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") pod \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\" (UID: \"cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.404224 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") pod \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\" (UID: \"a0128f3a-b052-44ed-a84e-c4c8aaf17c13\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.404240 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") pod \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\" (UID: \"bc5039c0-ea34-426b-a2b7-fbbc87b49a6d\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.404257 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") pod \"01ab3dd5-8196-46d0-ad33-122e2ca51def\" (UID: \"01ab3dd5-8196-46d0-ad33-122e2ca51def\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.404274 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.404289 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") pod \"7539238d-5fe0-46ed-884e-1c3b566537ec\" (UID: \"7539238d-5fe0-46ed-884e-1c3b566537ec\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.404319 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.404336 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.404352 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.404367 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.404381 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.404397 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.404413 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.404429 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") pod \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\" (UID: \"b6cd30de-2eeb-49a2-ab40-9167f4560ff5\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.404446 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") pod \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\" (UID: \"bd23aa5c-e532-4e53-bccf-e79f130c5ae8\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.404460 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.404477 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") pod \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\" (UID: \"b6312bbd-5731-4ea0-a20f-81d5a57df44a\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.404493 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.404514 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.404537 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.404574 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.404731 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.404819 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") pod \"6ea678ab-3438-413e-bfe3-290ae7725660\" (UID: \"6ea678ab-3438-413e-bfe3-290ae7725660\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.404865 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.404900 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.404936 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.404972 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.405009 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.405046 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.405079 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") pod \"57a731c4-ef35-47a8-b875-bfb08a7f8011\" (UID: \"57a731c4-ef35-47a8-b875-bfb08a7f8011\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.405117 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") pod \"9d4552c7-cd75-42dd-8880-30dd377c49a4\" (UID: \"9d4552c7-cd75-42dd-8880-30dd377c49a4\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.405147 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.405178 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.405212 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.405251 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") pod \"7583ce53-e0fe-4a16-9e4d-50516596a136\" (UID: \"7583ce53-e0fe-4a16-9e4d-50516596a136\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.405291 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") pod \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\" (UID: \"308be0ea-9f5f-4b29-aeb1-5abd31a0b17b\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.405372 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") pod \"fda69060-fa79-4696-b1a6-7980f124bf7c\" (UID: \"fda69060-fa79-4696-b1a6-7980f124bf7c\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.405404 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") pod \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\" (UID: \"1386a44e-36a2-460c-96d0-0359d2b6f0f5\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.405435 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.405465 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") pod \"1bf7eb37-55a3-4c65-b768-a94c82151e69\" (UID: \"1bf7eb37-55a3-4c65-b768-a94c82151e69\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.405496 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.405527 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.405561 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") pod \"20b0d48f-5fd6-431c-a545-e3c800c7b866\" (UID: \"20b0d48f-5fd6-431c-a545-e3c800c7b866\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.405596 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.405629 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.405660 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.405694 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") pod \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\" (UID: \"3ab1a177-2de0-46d9-b765-d0d0649bb42e\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.405729 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") pod \"1d611f23-29be-4491-8495-bee1670e935f\" (UID: \"1d611f23-29be-4491-8495-bee1670e935f\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.405762 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") pod \"6509e943-70c6-444c-bc41-48a544e36fbd\" (UID: \"6509e943-70c6-444c-bc41-48a544e36fbd\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.405795 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.405827 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.405925 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.405968 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") pod \"7bb08738-c794-4ee8-9972-3a62ca171029\" (UID: \"7bb08738-c794-4ee8-9972-3a62ca171029\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.406000 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") pod \"09efc573-dbb6-4249-bd59-9b87aba8dd28\" (UID: \"09efc573-dbb6-4249-bd59-9b87aba8dd28\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.406021 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") pod \"31d8b7a1-420e-4252-a5b7-eebe8a111292\" (UID: \"31d8b7a1-420e-4252-a5b7-eebe8a111292\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.406047 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.406073 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") pod \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\" (UID: \"25e176fe-21b4-4974-b1ed-c8b94f112a7f\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.406097 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") pod \"5441d097-087c-4d9a-baa8-b210afa90fc9\" (UID: \"5441d097-087c-4d9a-baa8-b210afa90fc9\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.406203 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") pod \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\" (UID: \"49c341d1-5089-4bc2-86a0-a5e165cfcc6b\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.406230 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") pod \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\" (UID: \"09ae3b1a-e8e7-4524-b54b-61eab6f9239a\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.406254 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") pod \"43509403-f426-496e-be36-56cef71462f5\" (UID: \"43509403-f426-496e-be36-56cef71462f5\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.406279 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") pod \"87cf06ed-a83f-41a7-828d-70653580a8cb\" (UID: \"87cf06ed-a83f-41a7-828d-70653580a8cb\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.406323 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.406358 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.406383 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") pod \"22c825df-677d-4ca6-82db-3454ed06e783\" (UID: \"22c825df-677d-4ca6-82db-3454ed06e783\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.406405 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") pod \"0b574797-001e-440a-8f4e-c0be86edad0f\" (UID: \"0b574797-001e-440a-8f4e-c0be86edad0f\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.406485 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-var-lib-openvswitch\") pod \"ovnkube-node-f7mp5\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.406510 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7e483c2b-08f0-4e92-8e4a-b7281f30af3e-proxy-tls\") pod \"machine-config-daemon-m4xwc\" (UID: \"7e483c2b-08f0-4e92-8e4a-b7281f30af3e\") " pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.406529 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/3b7580c2-5c23-4c67-807a-ea97a3df9398-cnibin\") pod \"multus-bwpqp\" (UID: \"3b7580c2-5c23-4c67-807a-ea97a3df9398\") " pod="openshift-multus/multus-bwpqp" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.406549 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/15117613-9fad-48c7-98c4-a2d84502ded9-ovnkube-config\") pod \"ovnkube-node-f7mp5\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.406575 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hjpwd\" (UniqueName: \"kubernetes.io/projected/5b5233d6-d2b8-42dc-b881-18e439ecc442-kube-api-access-hjpwd\") pod \"ovnkube-control-plane-749d76644c-4xdd8\" (UID: \"5b5233d6-d2b8-42dc-b881-18e439ecc442\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xdd8" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.406595 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/3b7580c2-5c23-4c67-807a-ea97a3df9398-hostroot\") pod \"multus-bwpqp\" (UID: \"3b7580c2-5c23-4c67-807a-ea97a3df9398\") " pod="openshift-multus/multus-bwpqp" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.406617 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/b9f9c7f1-6593-4d92-854c-e6c9964a0435-hosts-file\") pod \"node-resolver-9s5fr\" (UID: \"b9f9c7f1-6593-4d92-854c-e6c9964a0435\") " pod="openshift-dns/node-resolver-9s5fr" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.406647 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.406669 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/3b7580c2-5c23-4c67-807a-ea97a3df9398-multus-cni-dir\") pod \"multus-bwpqp\" (UID: \"3b7580c2-5c23-4c67-807a-ea97a3df9398\") " pod="openshift-multus/multus-bwpqp" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.406690 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/3b7580c2-5c23-4c67-807a-ea97a3df9398-host-run-k8s-cni-cncf-io\") pod \"multus-bwpqp\" (UID: \"3b7580c2-5c23-4c67-807a-ea97a3df9398\") " pod="openshift-multus/multus-bwpqp" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.406712 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.408908 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.409266 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/662dcbe7-7cf8-42d4-9a5e-345c32a49972-cnibin\") pod \"multus-additional-cni-plugins-mfp2k\" (UID: \"662dcbe7-7cf8-42d4-9a5e-345c32a49972\") " pod="openshift-multus/multus-additional-cni-plugins-mfp2k" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.409409 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/662dcbe7-7cf8-42d4-9a5e-345c32a49972-os-release\") pod \"multus-additional-cni-plugins-mfp2k\" (UID: \"662dcbe7-7cf8-42d4-9a5e-345c32a49972\") " pod="openshift-multus/multus-additional-cni-plugins-mfp2k" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.409516 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/662dcbe7-7cf8-42d4-9a5e-345c32a49972-cni-binary-copy\") pod \"multus-additional-cni-plugins-mfp2k\" (UID: \"662dcbe7-7cf8-42d4-9a5e-345c32a49972\") " pod="openshift-multus/multus-additional-cni-plugins-mfp2k" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.409682 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/3b7580c2-5c23-4c67-807a-ea97a3df9398-multus-socket-dir-parent\") pod \"multus-bwpqp\" (UID: \"3b7580c2-5c23-4c67-807a-ea97a3df9398\") " pod="openshift-multus/multus-bwpqp" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.409841 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-run-openvswitch\") pod \"ovnkube-node-f7mp5\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.409956 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/15117613-9fad-48c7-98c4-a2d84502ded9-env-overrides\") pod \"ovnkube-node-f7mp5\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.410058 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.410225 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.410384 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5b5233d6-d2b8-42dc-b881-18e439ecc442-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-4xdd8\" (UID: \"5b5233d6-d2b8-42dc-b881-18e439ecc442\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xdd8" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.410781 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.410921 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3b7580c2-5c23-4c67-807a-ea97a3df9398-host-run-netns\") pod \"multus-bwpqp\" (UID: \"3b7580c2-5c23-4c67-807a-ea97a3df9398\") " pod="openshift-multus/multus-bwpqp" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.411057 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/662dcbe7-7cf8-42d4-9a5e-345c32a49972-tuning-conf-dir\") pod \"multus-additional-cni-plugins-mfp2k\" (UID: \"662dcbe7-7cf8-42d4-9a5e-345c32a49972\") " pod="openshift-multus/multus-additional-cni-plugins-mfp2k" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.411163 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/d0d6440c-4524-42ea-b18a-0ddc4662f672-serviceca\") pod \"node-ca-njc9c\" (UID: \"d0d6440c-4524-42ea-b18a-0ddc4662f672\") " pod="openshift-image-registry/node-ca-njc9c" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.411263 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-run-systemd\") pod \"ovnkube-node-f7mp5\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.411383 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-host-run-ovn-kubernetes\") pod \"ovnkube-node-f7mp5\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.411472 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-f7mp5\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.411549 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5b5233d6-d2b8-42dc-b881-18e439ecc442-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-4xdd8\" (UID: \"5b5233d6-d2b8-42dc-b881-18e439ecc442\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xdd8" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.411616 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-host-kubelet\") pod \"ovnkube-node-f7mp5\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.411683 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-run-ovn\") pod \"ovnkube-node-f7mp5\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.411763 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/3b7580c2-5c23-4c67-807a-ea97a3df9398-os-release\") pod \"multus-bwpqp\" (UID: \"3b7580c2-5c23-4c67-807a-ea97a3df9398\") " pod="openshift-multus/multus-bwpqp" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.411864 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/3b7580c2-5c23-4c67-807a-ea97a3df9398-multus-conf-dir\") pod \"multus-bwpqp\" (UID: \"3b7580c2-5c23-4c67-807a-ea97a3df9398\") " pod="openshift-multus/multus-bwpqp" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.411958 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-etc-openvswitch\") pod \"ovnkube-node-f7mp5\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.412054 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-log-socket\") pod \"ovnkube-node-f7mp5\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.412147 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/15117613-9fad-48c7-98c4-a2d84502ded9-ovn-node-metrics-cert\") pod \"ovnkube-node-f7mp5\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.412250 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/15117613-9fad-48c7-98c4-a2d84502ded9-ovnkube-script-lib\") pod \"ovnkube-node-f7mp5\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.412381 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/be985ff1-de8a-431f-a230-67894ebe52d5-metrics-certs\") pod \"network-metrics-daemon-mqmtd\" (UID: \"be985ff1-de8a-431f-a230-67894ebe52d5\") " pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.412475 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5b5233d6-d2b8-42dc-b881-18e439ecc442-env-overrides\") pod \"ovnkube-control-plane-749d76644c-4xdd8\" (UID: \"5b5233d6-d2b8-42dc-b881-18e439ecc442\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xdd8" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.412585 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/3b7580c2-5c23-4c67-807a-ea97a3df9398-host-var-lib-kubelet\") pod \"multus-bwpqp\" (UID: \"3b7580c2-5c23-4c67-807a-ea97a3df9398\") " pod="openshift-multus/multus-bwpqp" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.412682 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/3b7580c2-5c23-4c67-807a-ea97a3df9398-host-run-multus-certs\") pod \"multus-bwpqp\" (UID: \"3b7580c2-5c23-4c67-807a-ea97a3df9398\") " pod="openshift-multus/multus-bwpqp" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.412768 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s8lxm\" (UniqueName: \"kubernetes.io/projected/3b7580c2-5c23-4c67-807a-ea97a3df9398-kube-api-access-s8lxm\") pod \"multus-bwpqp\" (UID: \"3b7580c2-5c23-4c67-807a-ea97a3df9398\") " pod="openshift-multus/multus-bwpqp" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.412851 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.412936 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/3b7580c2-5c23-4c67-807a-ea97a3df9398-cni-binary-copy\") pod \"multus-bwpqp\" (UID: \"3b7580c2-5c23-4c67-807a-ea97a3df9398\") " pod="openshift-multus/multus-bwpqp" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.413074 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3b7580c2-5c23-4c67-807a-ea97a3df9398-host-var-lib-cni-bin\") pod \"multus-bwpqp\" (UID: \"3b7580c2-5c23-4c67-807a-ea97a3df9398\") " pod="openshift-multus/multus-bwpqp" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.413246 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.413357 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/662dcbe7-7cf8-42d4-9a5e-345c32a49972-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-mfp2k\" (UID: \"662dcbe7-7cf8-42d4-9a5e-345c32a49972\") " pod="openshift-multus/multus-additional-cni-plugins-mfp2k" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.413450 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b6psg\" (UniqueName: \"kubernetes.io/projected/662dcbe7-7cf8-42d4-9a5e-345c32a49972-kube-api-access-b6psg\") pod \"multus-additional-cni-plugins-mfp2k\" (UID: \"662dcbe7-7cf8-42d4-9a5e-345c32a49972\") " pod="openshift-multus/multus-additional-cni-plugins-mfp2k" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.413535 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gd9jv\" (UniqueName: \"kubernetes.io/projected/7e483c2b-08f0-4e92-8e4a-b7281f30af3e-kube-api-access-gd9jv\") pod \"machine-config-daemon-m4xwc\" (UID: \"7e483c2b-08f0-4e92-8e4a-b7281f30af3e\") " pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.413857 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-host-slash\") pod \"ovnkube-node-f7mp5\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.413965 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-host-cni-bin\") pod \"ovnkube-node-f7mp5\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.414052 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dvwhv\" (UniqueName: \"kubernetes.io/projected/15117613-9fad-48c7-98c4-a2d84502ded9-kube-api-access-dvwhv\") pod \"ovnkube-node-f7mp5\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.414143 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.414235 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/7e483c2b-08f0-4e92-8e4a-b7281f30af3e-rootfs\") pod \"machine-config-daemon-m4xwc\" (UID: \"7e483c2b-08f0-4e92-8e4a-b7281f30af3e\") " pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.414344 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.414430 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/3b7580c2-5c23-4c67-807a-ea97a3df9398-system-cni-dir\") pod \"multus-bwpqp\" (UID: \"3b7580c2-5c23-4c67-807a-ea97a3df9398\") " pod="openshift-multus/multus-bwpqp" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.414529 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.416153 4948 swap_util.go:74] "error creating dir to test if tmpfs noswap is enabled. Assuming not supported" mount path="" error="stat /var/lib/kubelet/plugins/kubernetes.io/empty-dir: no such file or directory" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.416516 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"iptables-alerter-script\" (UniqueName: \"kubernetes.io/configmap/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-iptables-alerter-script\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.416520 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.416603 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hmsz5\" (UniqueName: \"kubernetes.io/projected/be985ff1-de8a-431f-a230-67894ebe52d5-kube-api-access-hmsz5\") pod \"network-metrics-daemon-mqmtd\" (UID: \"be985ff1-de8a-431f-a230-67894ebe52d5\") " pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.416622 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/3b7580c2-5c23-4c67-807a-ea97a3df9398-host-var-lib-cni-multus\") pod \"multus-bwpqp\" (UID: \"3b7580c2-5c23-4c67-807a-ea97a3df9398\") " pod="openshift-multus/multus-bwpqp" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.416642 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3b7580c2-5c23-4c67-807a-ea97a3df9398-etc-kubernetes\") pod \"multus-bwpqp\" (UID: \"3b7580c2-5c23-4c67-807a-ea97a3df9398\") " pod="openshift-multus/multus-bwpqp" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.416687 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9xf4z\" (UniqueName: \"kubernetes.io/projected/b9f9c7f1-6593-4d92-854c-e6c9964a0435-kube-api-access-9xf4z\") pod \"node-resolver-9s5fr\" (UID: \"b9f9c7f1-6593-4d92-854c-e6c9964a0435\") " pod="openshift-dns/node-resolver-9s5fr" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.416712 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-node-log\") pod \"ovnkube-node-f7mp5\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.416729 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-host-cni-netd\") pod \"ovnkube-node-f7mp5\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.416755 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.416778 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-host-run-netns\") pod \"ovnkube-node-f7mp5\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.416803 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7e483c2b-08f0-4e92-8e4a-b7281f30af3e-mcd-auth-proxy-config\") pod \"machine-config-daemon-m4xwc\" (UID: \"7e483c2b-08f0-4e92-8e4a-b7281f30af3e\") " pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.416827 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/3b7580c2-5c23-4c67-807a-ea97a3df9398-multus-daemon-config\") pod \"multus-bwpqp\" (UID: \"3b7580c2-5c23-4c67-807a-ea97a3df9398\") " pod="openshift-multus/multus-bwpqp" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.416849 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/662dcbe7-7cf8-42d4-9a5e-345c32a49972-system-cni-dir\") pod \"multus-additional-cni-plugins-mfp2k\" (UID: \"662dcbe7-7cf8-42d4-9a5e-345c32a49972\") " pod="openshift-multus/multus-additional-cni-plugins-mfp2k" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.416872 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k7wn2\" (UniqueName: \"kubernetes.io/projected/d0d6440c-4524-42ea-b18a-0ddc4662f672-kube-api-access-k7wn2\") pod \"node-ca-njc9c\" (UID: \"d0d6440c-4524-42ea-b18a-0ddc4662f672\") " pod="openshift-image-registry/node-ca-njc9c" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.416894 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-systemd-units\") pod \"ovnkube-node-f7mp5\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.416924 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.416947 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d0d6440c-4524-42ea-b18a-0ddc4662f672-host\") pod \"node-ca-njc9c\" (UID: \"d0d6440c-4524-42ea-b18a-0ddc4662f672\") " pod="openshift-image-registry/node-ca-njc9c" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.417164 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.417927 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-env-overrides\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.422698 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.422811 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.422879 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.422938 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.422996 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:14Z","lastTransitionTime":"2026-03-12T00:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.425165 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/37a5e44f-9a88-4405-be8a-b645485e7312-metrics-tls\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.406097 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.407198 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates" (OuterVolumeSpecName: "available-featuregates") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "available-featuregates". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.408182 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb" (OuterVolumeSpecName: "kube-api-access-279lb") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "kube-api-access-279lb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.408217 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images" (OuterVolumeSpecName: "images") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.408329 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb" (OuterVolumeSpecName: "kube-api-access-mg5zb") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "kube-api-access-mg5zb". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.408541 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c" (OuterVolumeSpecName: "kube-api-access-tk88c") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "kube-api-access-tk88c". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.408669 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images" (OuterVolumeSpecName: "images") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "images". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.409932 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.409861 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca" (OuterVolumeSpecName: "service-ca") pod "0b78653f-4ff9-4508-8672-245ed9b561e3" (UID: "0b78653f-4ff9-4508-8672-245ed9b561e3"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.409966 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs" (OuterVolumeSpecName: "webhook-certs") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "webhook-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.409998 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth" (OuterVolumeSpecName: "stats-auth") pod "c03ee662-fb2f-4fc4-a2c1-af487c19d254" (UID: "c03ee662-fb2f-4fc4-a2c1-af487c19d254"). InnerVolumeSpecName "stats-auth". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.410037 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs" (OuterVolumeSpecName: "kube-api-access-pcxfs") pod "9d4552c7-cd75-42dd-8880-30dd377c49a4" (UID: "9d4552c7-cd75-42dd-8880-30dd377c49a4"). InnerVolumeSpecName "kube-api-access-pcxfs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.410253 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls" (OuterVolumeSpecName: "samples-operator-tls") pod "a0128f3a-b052-44ed-a84e-c4c8aaf17c13" (UID: "a0128f3a-b052-44ed-a84e-c4c8aaf17c13"). InnerVolumeSpecName "samples-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.410388 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "496e6271-fb68-4057-954e-a0d97a4afa3f" (UID: "496e6271-fb68-4057-954e-a0d97a4afa3f"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.410612 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.410718 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp" (OuterVolumeSpecName: "kube-api-access-fcqwp") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "kube-api-access-fcqwp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.410723 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8" (OuterVolumeSpecName: "kube-api-access-6ccd8") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "kube-api-access-6ccd8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.410793 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume" (OuterVolumeSpecName: "config-volume") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.410998 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.411024 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.411274 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8" (OuterVolumeSpecName: "kube-api-access-wxkg8") pod "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" (UID: "3cb93b32-e0ae-4377-b9c8-fdb9842c6d59"). InnerVolumeSpecName "kube-api-access-wxkg8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.411324 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz" (OuterVolumeSpecName: "kube-api-access-8tdtz") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "kube-api-access-8tdtz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.411761 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7539238d-5fe0-46ed-884e-1c3b566537ec" (UID: "7539238d-5fe0-46ed-884e-1c3b566537ec"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.411886 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.411921 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.412051 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7" (OuterVolumeSpecName: "kube-api-access-9xfj7") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "kube-api-access-9xfj7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.412061 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs" (OuterVolumeSpecName: "certs") pod "5fe579f8-e8a6-4643-bce5-a661393c4dde" (UID: "5fe579f8-e8a6-4643-bce5-a661393c4dde"). InnerVolumeSpecName "certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.412193 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn" (OuterVolumeSpecName: "kube-api-access-jkwtn") pod "5b88f790-22fa-440e-b583-365168c0b23d" (UID: "5b88f790-22fa-440e-b583-365168c0b23d"). InnerVolumeSpecName "kube-api-access-jkwtn". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.412261 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.412605 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config" (OuterVolumeSpecName: "mcc-auth-proxy-config") pod "0b574797-001e-440a-8f4e-c0be86edad0f" (UID: "0b574797-001e-440a-8f4e-c0be86edad0f"). InnerVolumeSpecName "mcc-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.409610 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca" (OuterVolumeSpecName: "etcd-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.412940 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert" (OuterVolumeSpecName: "webhook-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "webhook-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.413018 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config" (OuterVolumeSpecName: "multus-daemon-config") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "multus-daemon-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.413050 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk" (OuterVolumeSpecName: "kube-api-access-rnphk") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "kube-api-access-rnphk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.413047 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca" (OuterVolumeSpecName: "client-ca") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.413276 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls" (OuterVolumeSpecName: "machine-api-operator-tls") pod "6402fda4-df10-493c-b4e5-d0569419652d" (UID: "6402fda4-df10-493c-b4e5-d0569419652d"). InnerVolumeSpecName "machine-api-operator-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.413286 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.413395 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert" (OuterVolumeSpecName: "package-server-manager-serving-cert") pod "3ab1a177-2de0-46d9-b765-d0d0649bb42e" (UID: "3ab1a177-2de0-46d9-b765-d0d0649bb42e"). InnerVolumeSpecName "package-server-manager-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.413411 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds" (OuterVolumeSpecName: "kube-api-access-w9rds") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "kube-api-access-w9rds". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.413580 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz" (OuterVolumeSpecName: "kube-api-access-bf2bz") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "kube-api-access-bf2bz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.412080 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv" (OuterVolumeSpecName: "kube-api-access-d4lsv") pod "25e176fe-21b4-4974-b1ed-c8b94f112a7f" (UID: "25e176fe-21b4-4974-b1ed-c8b94f112a7f"). InnerVolumeSpecName "kube-api-access-d4lsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.413810 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "4bb40260-dbaa-4fb0-84df-5e680505d512" (UID: "4bb40260-dbaa-4fb0-84df-5e680505d512"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.413819 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" (UID: "8cea82b4-6893-4ddc-af9f-1bb5ae425c5b"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: E0312 00:08:14.413913 4948 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 12 00:08:14 crc kubenswrapper[4948]: E0312 00:08:14.431033 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-12 00:08:14.931011797 +0000 UTC m=+94.386615535 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.414328 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz" (OuterVolumeSpecName: "kube-api-access-2d4wz") pod "5441d097-087c-4d9a-baa8-b210afa90fc9" (UID: "5441d097-087c-4d9a-baa8-b210afa90fc9"). InnerVolumeSpecName "kube-api-access-2d4wz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.414619 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config" (OuterVolumeSpecName: "config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.414912 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.415062 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle" (OuterVolumeSpecName: "service-ca-bundle") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "service-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.416438 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.417699 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config" (OuterVolumeSpecName: "mcd-auth-proxy-config") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "mcd-auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.417915 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.418523 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls" (OuterVolumeSpecName: "metrics-tls") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "metrics-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.418610 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg" (OuterVolumeSpecName: "kube-api-access-dbsvg") pod "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" (UID: "f88749ec-7931-4ee7-b3fc-1ec5e11f92e9"). InnerVolumeSpecName "kube-api-access-dbsvg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.419006 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.419592 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85" (OuterVolumeSpecName: "kube-api-access-x2m85") pod "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" (UID: "cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d"). InnerVolumeSpecName "kube-api-access-x2m85". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.419782 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr" (OuterVolumeSpecName: "kube-api-access-249nr") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "kube-api-access-249nr". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.420689 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities" (OuterVolumeSpecName: "utilities") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.420705 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.432354 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2w9zh\" (UniqueName: \"kubernetes.io/projected/4bb40260-dbaa-4fb0-84df-5e680505d512-kube-api-access-2w9zh\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.420927 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.420944 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2" (OuterVolumeSpecName: "kube-api-access-jhbk2") pod "bd23aa5c-e532-4e53-bccf-e79f130c5ae8" (UID: "bd23aa5c-e532-4e53-bccf-e79f130c5ae8"). InnerVolumeSpecName "kube-api-access-jhbk2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.421176 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert" (OuterVolumeSpecName: "cert") pod "20b0d48f-5fd6-431c-a545-e3c800c7b866" (UID: "20b0d48f-5fd6-431c-a545-e3c800c7b866"). InnerVolumeSpecName "cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.421377 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert" (OuterVolumeSpecName: "profile-collector-cert") pod "b6312bbd-5731-4ea0-a20f-81d5a57df44a" (UID: "b6312bbd-5731-4ea0-a20f-81d5a57df44a"). InnerVolumeSpecName "profile-collector-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.421583 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.423738 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.425282 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx" (OuterVolumeSpecName: "kube-api-access-d6qdx") pod "87cf06ed-a83f-41a7-828d-70653580a8cb" (UID: "87cf06ed-a83f-41a7-828d-70653580a8cb"). InnerVolumeSpecName "kube-api-access-d6qdx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.425358 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.425565 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.425889 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.426102 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert" (OuterVolumeSpecName: "apiservice-cert") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "apiservice-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.426365 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.426667 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt" (OuterVolumeSpecName: "kube-api-access-fqsjt") pod "efdd0498-1daa-4136-9a4a-3b948c2293fc" (UID: "efdd0498-1daa-4136-9a4a-3b948c2293fc"). InnerVolumeSpecName "kube-api-access-fqsjt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.426884 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.426777 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh" (OuterVolumeSpecName: "kube-api-access-xcgwh") pod "fda69060-fa79-4696-b1a6-7980f124bf7c" (UID: "fda69060-fa79-4696-b1a6-7980f124bf7c"). InnerVolumeSpecName "kube-api-access-xcgwh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: E0312 00:08:14.428750 4948 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.429095 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.432786 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6g6sz\" (UniqueName: \"kubernetes.io/projected/6509e943-70c6-444c-bc41-48a544e36fbd-kube-api-access-6g6sz\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: E0312 00:08:14.432793 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-12 00:08:14.932776151 +0000 UTC m=+94.388379889 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.432823 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x4zgh\" (UniqueName: \"kubernetes.io/projected/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-kube-api-access-x4zgh\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.432858 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-v47cf\" (UniqueName: \"kubernetes.io/projected/c03ee662-fb2f-4fc4-a2c1-af487c19d254-kube-api-access-v47cf\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.432905 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/e7e6199b-1264-4501-8953-767f51328d08-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.432918 4948 reconciler_common.go:293] "Volume detached for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-key\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.432930 4948 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.432939 4948 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.432951 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/496e6271-fb68-4057-954e-a0d97a4afa3f-config\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.432961 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qs4fp\" (UniqueName: \"kubernetes.io/projected/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-kube-api-access-qs4fp\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.432971 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w4xd4\" (UniqueName: \"kubernetes.io/projected/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-kube-api-access-w4xd4\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.432981 4948 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/5b88f790-22fa-440e-b583-365168c0b23d-metrics-certs\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.432993 4948 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/a31745f5-9847-4afe-82a5-3161cc66ca93-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.433003 4948 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.433012 4948 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/a31745f5-9847-4afe-82a5-3161cc66ca93-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.433021 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x7zkh\" (UniqueName: \"kubernetes.io/projected/6731426b-95fe-49ff-bb5f-40441049fde2-kube-api-access-x7zkh\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.433031 4948 reconciler_common.go:293] "Volume detached for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/25e176fe-21b4-4974-b1ed-c8b94f112a7f-signing-cabundle\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.433040 4948 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.433049 4948 reconciler_common.go:293] "Volume detached for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-audit\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.433058 4948 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/0b78653f-4ff9-4508-8672-245ed9b561e3-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.433067 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-sb6h7\" (UniqueName: \"kubernetes.io/projected/1bf7eb37-55a3-4c65-b768-a94c82151e69-kube-api-access-sb6h7\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.433076 4948 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.433085 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9d4552c7-cd75-42dd-8880-30dd377c49a4-config\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.433095 4948 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.433105 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.433115 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcphl\" (UniqueName: \"kubernetes.io/projected/7583ce53-e0fe-4a16-9e4d-50516596a136-kube-api-access-xcphl\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.433125 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-nzwt7\" (UniqueName: \"kubernetes.io/projected/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-kube-api-access-nzwt7\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.433135 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zkvpv\" (UniqueName: \"kubernetes.io/projected/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-kube-api-access-zkvpv\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.433144 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7539238d-5fe0-46ed-884e-1c3b566537ec-config\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.433153 4948 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.433163 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cfbct\" (UniqueName: \"kubernetes.io/projected/57a731c4-ef35-47a8-b875-bfb08a7f8011-kube-api-access-cfbct\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.433172 4948 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.433180 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-config\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.433190 4948 reconciler_common.go:293] "Volume detached for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/6731426b-95fe-49ff-bb5f-40441049fde2-control-plane-machine-set-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.433199 4948 reconciler_common.go:293] "Volume detached for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-image-import-ca\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.433209 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-vt5rc\" (UniqueName: \"kubernetes.io/projected/44663579-783b-4372-86d6-acf235a62d72-kube-api-access-vt5rc\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.433220 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gf66m\" (UniqueName: \"kubernetes.io/projected/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-kube-api-access-gf66m\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.433230 4948 reconciler_common.go:293] "Volume detached for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-console-config\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.433239 4948 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9d4552c7-cd75-42dd-8880-30dd377c49a4-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.433247 4948 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/c03ee662-fb2f-4fc4-a2c1-af487c19d254-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.433256 4948 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/96b93a3a-6083-4aea-8eab-fe1aa8245ad9-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.433265 4948 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.433276 4948 reconciler_common.go:293] "Volume detached for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf126b07-da06-4140-9a57-dfd54fc6b486-image-registry-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.433285 4948 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5441d097-087c-4d9a-baa8-b210afa90fc9-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.433295 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.433317 4948 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-client-ca\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.433326 4948 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.433336 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4d4hj\" (UniqueName: \"kubernetes.io/projected/3ab1a177-2de0-46d9-b765-d0d0649bb42e-kube-api-access-4d4hj\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.433346 4948 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-srv-cert\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.433355 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lz9wn\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-kube-api-access-lz9wn\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.433365 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-config\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.433372 4948 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-serviceca\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.433381 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-config\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.433389 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-config\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.433397 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-config\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.433406 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s4n52\" (UniqueName: \"kubernetes.io/projected/925f1c65-6136-48ba-85aa-3a3b50560753-kube-api-access-s4n52\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.433415 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/496e6271-fb68-4057-954e-a0d97a4afa3f-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.433425 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lzf88\" (UniqueName: \"kubernetes.io/projected/0b574797-001e-440a-8f4e-c0be86edad0f-kube-api-access-lzf88\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.433434 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/1386a44e-36a2-460c-96d0-0359d2b6f0f5-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.433443 4948 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/925f1c65-6136-48ba-85aa-3a3b50560753-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.433451 4948 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.433461 4948 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/e7e6199b-1264-4501-8953-767f51328d08-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.433469 4948 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.433478 4948 reconciler_common.go:293] "Volume detached for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-metrics-certs\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.433487 4948 reconciler_common.go:293] "Volume detached for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/925f1c65-6136-48ba-85aa-3a3b50560753-ovn-control-plane-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.433500 4948 reconciler_common.go:293] "Volume detached for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.433509 4948 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6509e943-70c6-444c-bc41-48a544e36fbd-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.433518 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/0b78653f-4ff9-4508-8672-245ed9b561e3-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.433526 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-config\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.433535 4948 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.433543 4948 reconciler_common.go:293] "Volume detached for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-default-certificate\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.433552 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pjr6v\" (UniqueName: \"kubernetes.io/projected/49ef4625-1d3a-4a9f-b595-c2433d32326d-kube-api-access-pjr6v\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.433561 4948 reconciler_common.go:293] "Volume detached for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-srv-cert\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.433569 4948 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.433577 4948 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.433585 4948 reconciler_common.go:293] "Volume detached for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-node-bootstrap-token\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.433593 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/e7e6199b-1264-4501-8953-767f51328d08-config\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.433602 4948 reconciler_common.go:293] "Volume detached for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/43509403-f426-496e-be36-56cef71462f5-console-oauth-config\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.434153 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: E0312 00:08:14.434543 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 12 00:08:14 crc kubenswrapper[4948]: E0312 00:08:14.434567 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 12 00:08:14 crc kubenswrapper[4948]: E0312 00:08:14.434583 4948 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 00:08:14 crc kubenswrapper[4948]: E0312 00:08:14.434634 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-12 00:08:14.934620179 +0000 UTC m=+94.390224007 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.436975 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca" (OuterVolumeSpecName: "etcd-serving-ca") pod "09ae3b1a-e8e7-4524-b54b-61eab6f9239a" (UID: "09ae3b1a-e8e7-4524-b54b-61eab6f9239a"). InnerVolumeSpecName "etcd-serving-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: E0312 00:08:14.437281 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 12 00:08:14 crc kubenswrapper[4948]: E0312 00:08:14.437320 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 12 00:08:14 crc kubenswrapper[4948]: E0312 00:08:14.437332 4948 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 00:08:14 crc kubenswrapper[4948]: E0312 00:08:14.437373 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-12 00:08:14.937360319 +0000 UTC m=+94.392964167 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.437455 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.437855 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp" (OuterVolumeSpecName: "kube-api-access-ngvvp") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "kube-api-access-ngvvp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.438112 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.438112 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/ef543e1b-8068-4ea3-b32a-61027b32e95d-webhook-cert\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.438554 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client" (OuterVolumeSpecName: "etcd-client") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "etcd-client". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.438705 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2kz5\" (UniqueName: \"kubernetes.io/projected/ef543e1b-8068-4ea3-b32a-61027b32e95d-kube-api-access-s2kz5\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.438741 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config" (OuterVolumeSpecName: "encryption-config") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "encryption-config". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.439054 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd" (OuterVolumeSpecName: "kube-api-access-mnrrd") pod "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" (UID: "bc5039c0-ea34-426b-a2b7-fbbc87b49a6d"). InnerVolumeSpecName "kube-api-access-mnrrd". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.439183 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.439199 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls" (OuterVolumeSpecName: "proxy-tls") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "proxy-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.440123 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert" (OuterVolumeSpecName: "oauth-serving-cert") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "oauth-serving-cert". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.440132 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle" (OuterVolumeSpecName: "trusted-ca-bundle") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.440498 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rdwmf\" (UniqueName: \"kubernetes.io/projected/37a5e44f-9a88-4405-be8a-b645485e7312-kube-api-access-rdwmf\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.440680 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "1bf7eb37-55a3-4c65-b768-a94c82151e69" (UID: "1bf7eb37-55a3-4c65-b768-a94c82151e69"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.440769 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-identity-cm\" (UniqueName: \"kubernetes.io/configmap/ef543e1b-8068-4ea3-b32a-61027b32e95d-ovnkube-identity-cm\") pod \"network-node-identity-vrzqb\" (UID: \"ef543e1b-8068-4ea3-b32a-61027b32e95d\") " pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.441289 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5" (OuterVolumeSpecName: "kube-api-access-zgdk5") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "kube-api-access-zgdk5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.441492 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca" (OuterVolumeSpecName: "service-ca") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.442019 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "31d8b7a1-420e-4252-a5b7-eebe8a111292" (UID: "31d8b7a1-420e-4252-a5b7-eebe8a111292"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.444171 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.445667 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist" (OuterVolumeSpecName: "cni-sysctl-allowlist") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-sysctl-allowlist". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.446473 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities" (OuterVolumeSpecName: "utilities") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.446637 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6" (OuterVolumeSpecName: "kube-api-access-htfz6") pod "6ea678ab-3438-413e-bfe3-290ae7725660" (UID: "6ea678ab-3438-413e-bfe3-290ae7725660"). InnerVolumeSpecName "kube-api-access-htfz6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.446681 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "49c341d1-5089-4bc2-86a0-a5e165cfcc6b" (UID: "49c341d1-5089-4bc2-86a0-a5e165cfcc6b"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.446683 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "a31745f5-9847-4afe-82a5-3161cc66ca93" (UID: "a31745f5-9847-4afe-82a5-3161cc66ca93"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.447006 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls" (OuterVolumeSpecName: "machine-approver-tls") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "machine-approver-tls". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.447072 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca" (OuterVolumeSpecName: "etcd-service-ca") pod "09efc573-dbb6-4249-bd59-9b87aba8dd28" (UID: "09efc573-dbb6-4249-bd59-9b87aba8dd28"). InnerVolumeSpecName "etcd-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.447559 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf" (OuterVolumeSpecName: "kube-api-access-7c4vf") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "kube-api-access-7c4vf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.448854 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rczfb\" (UniqueName: \"kubernetes.io/projected/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-kube-api-access-rczfb\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.449055 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs" (OuterVolumeSpecName: "tmpfs") pod "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" (UID: "308be0ea-9f5f-4b29-aeb1-5abd31a0b17b"). InnerVolumeSpecName "tmpfs". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.449212 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "bf126b07-da06-4140-9a57-dfd54fc6b486" (UID: "bf126b07-da06-4140-9a57-dfd54fc6b486"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.449503 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config" (OuterVolumeSpecName: "config") pod "1386a44e-36a2-460c-96d0-0359d2b6f0f5" (UID: "1386a44e-36a2-460c-96d0-0359d2b6f0f5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.449871 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config" (OuterVolumeSpecName: "config") pod "6509e943-70c6-444c-bc41-48a544e36fbd" (UID: "6509e943-70c6-444c-bc41-48a544e36fbd"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.450025 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config" (OuterVolumeSpecName: "auth-proxy-config") pod "22c825df-677d-4ca6-82db-3454ed06e783" (UID: "22c825df-677d-4ca6-82db-3454ed06e783"). InnerVolumeSpecName "auth-proxy-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.450494 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7" (OuterVolumeSpecName: "kube-api-access-kfwg7") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "kube-api-access-kfwg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.450559 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy" (OuterVolumeSpecName: "cni-binary-copy") pod "7bb08738-c794-4ee8-9972-3a62ca171029" (UID: "7bb08738-c794-4ee8-9972-3a62ca171029"). InnerVolumeSpecName "cni-binary-copy". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.450566 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.451108 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.451197 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782" (OuterVolumeSpecName: "kube-api-access-pj782") pod "b6cd30de-2eeb-49a2-ab40-9167f4560ff5" (UID: "b6cd30de-2eeb-49a2-ab40-9167f4560ff5"). InnerVolumeSpecName "kube-api-access-pj782". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.451493 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config" (OuterVolumeSpecName: "config") pod "7583ce53-e0fe-4a16-9e4d-50516596a136" (UID: "7583ce53-e0fe-4a16-9e4d-50516596a136"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.451534 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mqmtd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be985ff1-de8a-431f-a230-67894ebe52d5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmsz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmsz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mqmtd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.451986 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "1d611f23-29be-4491-8495-bee1670e935f" (UID: "1d611f23-29be-4491-8495-bee1670e935f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.452110 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5" (OuterVolumeSpecName: "kube-api-access-qg5z5") pod "43509403-f426-496e-be36-56cef71462f5" (UID: "43509403-f426-496e-be36-56cef71462f5"). InnerVolumeSpecName "kube-api-access-qg5z5". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.453394 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j" (OuterVolumeSpecName: "kube-api-access-w7l8j") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "kube-api-access-w7l8j". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.453774 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "57a731c4-ef35-47a8-b875-bfb08a7f8011" (UID: "57a731c4-ef35-47a8-b875-bfb08a7f8011"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.454729 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config" (OuterVolumeSpecName: "config") pod "01ab3dd5-8196-46d0-ad33-122e2ca51def" (UID: "01ab3dd5-8196-46d0-ad33-122e2ca51def"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.455980 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" (UID: "b11524ee-3fca-4b1b-9cdf-6da289fdbc7d"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.462022 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.464809 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15117613-9fad-48c7-98c4-a2d84502ded9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7mp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.467535 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5225d0e4-402f-4861-b410-819f433b1803" (UID: "5225d0e4-402f-4861-b410-819f433b1803"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.472501 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.479670 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.487173 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.493952 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.508670 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfp2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"662dcbe7-7cf8-42d4-9a5e-345c32a49972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfp2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.516165 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e483c2b-08f0-4e92-8e4a-b7281f30af3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd9jv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd9jv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4xwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.524114 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-operator]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":false,\\\"restartCount\\\":5,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.528149 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.528200 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.528220 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.528237 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.528252 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:14Z","lastTransitionTime":"2026-03-12T00:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.533936 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/7e483c2b-08f0-4e92-8e4a-b7281f30af3e-rootfs\") pod \"machine-config-daemon-m4xwc\" (UID: \"7e483c2b-08f0-4e92-8e4a-b7281f30af3e\") " pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.533961 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gd9jv\" (UniqueName: \"kubernetes.io/projected/7e483c2b-08f0-4e92-8e4a-b7281f30af3e-kube-api-access-gd9jv\") pod \"machine-config-daemon-m4xwc\" (UID: \"7e483c2b-08f0-4e92-8e4a-b7281f30af3e\") " pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.534015 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"rootfs\" (UniqueName: \"kubernetes.io/host-path/7e483c2b-08f0-4e92-8e4a-b7281f30af3e-rootfs\") pod \"machine-config-daemon-m4xwc\" (UID: \"7e483c2b-08f0-4e92-8e4a-b7281f30af3e\") " pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.534051 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-host-slash\") pod \"ovnkube-node-f7mp5\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.534067 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-host-cni-bin\") pod \"ovnkube-node-f7mp5\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.534083 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dvwhv\" (UniqueName: \"kubernetes.io/projected/15117613-9fad-48c7-98c4-a2d84502ded9-kube-api-access-dvwhv\") pod \"ovnkube-node-f7mp5\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.534363 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/3b7580c2-5c23-4c67-807a-ea97a3df9398-system-cni-dir\") pod \"multus-bwpqp\" (UID: \"3b7580c2-5c23-4c67-807a-ea97a3df9398\") " pod="openshift-multus/multus-bwpqp" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.534429 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hmsz5\" (UniqueName: \"kubernetes.io/projected/be985ff1-de8a-431f-a230-67894ebe52d5-kube-api-access-hmsz5\") pod \"network-metrics-daemon-mqmtd\" (UID: \"be985ff1-de8a-431f-a230-67894ebe52d5\") " pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.534465 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/3b7580c2-5c23-4c67-807a-ea97a3df9398-host-var-lib-cni-multus\") pod \"multus-bwpqp\" (UID: \"3b7580c2-5c23-4c67-807a-ea97a3df9398\") " pod="openshift-multus/multus-bwpqp" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.534482 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3b7580c2-5c23-4c67-807a-ea97a3df9398-etc-kubernetes\") pod \"multus-bwpqp\" (UID: \"3b7580c2-5c23-4c67-807a-ea97a3df9398\") " pod="openshift-multus/multus-bwpqp" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.534156 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-host-slash\") pod \"ovnkube-node-f7mp5\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.534798 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/3b7580c2-5c23-4c67-807a-ea97a3df9398-system-cni-dir\") pod \"multus-bwpqp\" (UID: \"3b7580c2-5c23-4c67-807a-ea97a3df9398\") " pod="openshift-multus/multus-bwpqp" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.534846 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-multus\" (UniqueName: \"kubernetes.io/host-path/3b7580c2-5c23-4c67-807a-ea97a3df9398-host-var-lib-cni-multus\") pod \"multus-bwpqp\" (UID: \"3b7580c2-5c23-4c67-807a-ea97a3df9398\") " pod="openshift-multus/multus-bwpqp" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.534229 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-host-cni-bin\") pod \"ovnkube-node-f7mp5\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.534965 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-kubernetes\" (UniqueName: \"kubernetes.io/host-path/3b7580c2-5c23-4c67-807a-ea97a3df9398-etc-kubernetes\") pod \"multus-bwpqp\" (UID: \"3b7580c2-5c23-4c67-807a-ea97a3df9398\") " pod="openshift-multus/multus-bwpqp" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.535133 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9xf4z\" (UniqueName: \"kubernetes.io/projected/b9f9c7f1-6593-4d92-854c-e6c9964a0435-kube-api-access-9xf4z\") pod \"node-resolver-9s5fr\" (UID: \"b9f9c7f1-6593-4d92-854c-e6c9964a0435\") " pod="openshift-dns/node-resolver-9s5fr" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.535159 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-node-log\") pod \"ovnkube-node-f7mp5\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.535174 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-host-cni-netd\") pod \"ovnkube-node-f7mp5\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.535188 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-host-run-netns\") pod \"ovnkube-node-f7mp5\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.535203 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7e483c2b-08f0-4e92-8e4a-b7281f30af3e-mcd-auth-proxy-config\") pod \"machine-config-daemon-m4xwc\" (UID: \"7e483c2b-08f0-4e92-8e4a-b7281f30af3e\") " pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.535296 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-node-log\") pod \"ovnkube-node-f7mp5\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.535351 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/3b7580c2-5c23-4c67-807a-ea97a3df9398-multus-daemon-config\") pod \"multus-bwpqp\" (UID: \"3b7580c2-5c23-4c67-807a-ea97a3df9398\") " pod="openshift-multus/multus-bwpqp" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.535384 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-host-cni-netd\") pod \"ovnkube-node-f7mp5\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.535416 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/662dcbe7-7cf8-42d4-9a5e-345c32a49972-system-cni-dir\") pod \"multus-additional-cni-plugins-mfp2k\" (UID: \"662dcbe7-7cf8-42d4-9a5e-345c32a49972\") " pod="openshift-multus/multus-additional-cni-plugins-mfp2k" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.535452 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k7wn2\" (UniqueName: \"kubernetes.io/projected/d0d6440c-4524-42ea-b18a-0ddc4662f672-kube-api-access-k7wn2\") pod \"node-ca-njc9c\" (UID: \"d0d6440c-4524-42ea-b18a-0ddc4662f672\") " pod="openshift-image-registry/node-ca-njc9c" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.535486 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-systemd-units\") pod \"ovnkube-node-f7mp5\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.535536 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d0d6440c-4524-42ea-b18a-0ddc4662f672-host\") pod \"node-ca-njc9c\" (UID: \"d0d6440c-4524-42ea-b18a-0ddc4662f672\") " pod="openshift-image-registry/node-ca-njc9c" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.535575 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-var-lib-openvswitch\") pod \"ovnkube-node-f7mp5\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.535605 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7e483c2b-08f0-4e92-8e4a-b7281f30af3e-proxy-tls\") pod \"machine-config-daemon-m4xwc\" (UID: \"7e483c2b-08f0-4e92-8e4a-b7281f30af3e\") " pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.535634 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/3b7580c2-5c23-4c67-807a-ea97a3df9398-cnibin\") pod \"multus-bwpqp\" (UID: \"3b7580c2-5c23-4c67-807a-ea97a3df9398\") " pod="openshift-multus/multus-bwpqp" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.535662 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/15117613-9fad-48c7-98c4-a2d84502ded9-ovnkube-config\") pod \"ovnkube-node-f7mp5\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.535692 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hjpwd\" (UniqueName: \"kubernetes.io/projected/5b5233d6-d2b8-42dc-b881-18e439ecc442-kube-api-access-hjpwd\") pod \"ovnkube-control-plane-749d76644c-4xdd8\" (UID: \"5b5233d6-d2b8-42dc-b881-18e439ecc442\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xdd8" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.535724 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/3b7580c2-5c23-4c67-807a-ea97a3df9398-hostroot\") pod \"multus-bwpqp\" (UID: \"3b7580c2-5c23-4c67-807a-ea97a3df9398\") " pod="openshift-multus/multus-bwpqp" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.535758 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-systemd-units\") pod \"ovnkube-node-f7mp5\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.535828 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host\" (UniqueName: \"kubernetes.io/host-path/d0d6440c-4524-42ea-b18a-0ddc4662f672-host\") pod \"node-ca-njc9c\" (UID: \"d0d6440c-4524-42ea-b18a-0ddc4662f672\") " pod="openshift-image-registry/node-ca-njc9c" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.535853 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/b9f9c7f1-6593-4d92-854c-e6c9964a0435-hosts-file\") pod \"node-resolver-9s5fr\" (UID: \"b9f9c7f1-6593-4d92-854c-e6c9964a0435\") " pod="openshift-dns/node-resolver-9s5fr" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.535772 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"hosts-file\" (UniqueName: \"kubernetes.io/host-path/b9f9c7f1-6593-4d92-854c-e6c9964a0435-hosts-file\") pod \"node-resolver-9s5fr\" (UID: \"b9f9c7f1-6593-4d92-854c-e6c9964a0435\") " pod="openshift-dns/node-resolver-9s5fr" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.535953 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/3b7580c2-5c23-4c67-807a-ea97a3df9398-multus-cni-dir\") pod \"multus-bwpqp\" (UID: \"3b7580c2-5c23-4c67-807a-ea97a3df9398\") " pod="openshift-multus/multus-bwpqp" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.536001 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/3b7580c2-5c23-4c67-807a-ea97a3df9398-host-run-k8s-cni-cncf-io\") pod \"multus-bwpqp\" (UID: \"3b7580c2-5c23-4c67-807a-ea97a3df9398\") " pod="openshift-multus/multus-bwpqp" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.536022 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-cni-dir\" (UniqueName: \"kubernetes.io/host-path/3b7580c2-5c23-4c67-807a-ea97a3df9398-multus-cni-dir\") pod \"multus-bwpqp\" (UID: \"3b7580c2-5c23-4c67-807a-ea97a3df9398\") " pod="openshift-multus/multus-bwpqp" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.536035 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.536067 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/662dcbe7-7cf8-42d4-9a5e-345c32a49972-cnibin\") pod \"multus-additional-cni-plugins-mfp2k\" (UID: \"662dcbe7-7cf8-42d4-9a5e-345c32a49972\") " pod="openshift-multus/multus-additional-cni-plugins-mfp2k" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.536097 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/662dcbe7-7cf8-42d4-9a5e-345c32a49972-os-release\") pod \"multus-additional-cni-plugins-mfp2k\" (UID: \"662dcbe7-7cf8-42d4-9a5e-345c32a49972\") " pod="openshift-multus/multus-additional-cni-plugins-mfp2k" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.536129 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/662dcbe7-7cf8-42d4-9a5e-345c32a49972-cni-binary-copy\") pod \"multus-additional-cni-plugins-mfp2k\" (UID: \"662dcbe7-7cf8-42d4-9a5e-345c32a49972\") " pod="openshift-multus/multus-additional-cni-plugins-mfp2k" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.536147 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/3b7580c2-5c23-4c67-807a-ea97a3df9398-cnibin\") pod \"multus-bwpqp\" (UID: \"3b7580c2-5c23-4c67-807a-ea97a3df9398\") " pod="openshift-multus/multus-bwpqp" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.536161 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/3b7580c2-5c23-4c67-807a-ea97a3df9398-multus-socket-dir-parent\") pod \"multus-bwpqp\" (UID: \"3b7580c2-5c23-4c67-807a-ea97a3df9398\") " pod="openshift-multus/multus-bwpqp" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.536194 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-run-openvswitch\") pod \"ovnkube-node-f7mp5\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.536224 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/15117613-9fad-48c7-98c4-a2d84502ded9-env-overrides\") pod \"ovnkube-node-f7mp5\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.536268 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.536331 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5b5233d6-d2b8-42dc-b881-18e439ecc442-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-4xdd8\" (UID: \"5b5233d6-d2b8-42dc-b881-18e439ecc442\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xdd8" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.536363 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3b7580c2-5c23-4c67-807a-ea97a3df9398-host-run-netns\") pod \"multus-bwpqp\" (UID: \"3b7580c2-5c23-4c67-807a-ea97a3df9398\") " pod="openshift-multus/multus-bwpqp" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.536393 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/662dcbe7-7cf8-42d4-9a5e-345c32a49972-tuning-conf-dir\") pod \"multus-additional-cni-plugins-mfp2k\" (UID: \"662dcbe7-7cf8-42d4-9a5e-345c32a49972\") " pod="openshift-multus/multus-additional-cni-plugins-mfp2k" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.536425 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/d0d6440c-4524-42ea-b18a-0ddc4662f672-serviceca\") pod \"node-ca-njc9c\" (UID: \"d0d6440c-4524-42ea-b18a-0ddc4662f672\") " pod="openshift-image-registry/node-ca-njc9c" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.536456 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-run-systemd\") pod \"ovnkube-node-f7mp5\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.536519 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-host-run-ovn-kubernetes\") pod \"ovnkube-node-f7mp5\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.536528 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/15117613-9fad-48c7-98c4-a2d84502ded9-ovnkube-config\") pod \"ovnkube-node-f7mp5\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.536554 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-f7mp5\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.536605 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-f7mp5\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.536657 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/3b7580c2-5c23-4c67-807a-ea97a3df9398-host-run-netns\") pod \"multus-bwpqp\" (UID: \"3b7580c2-5c23-4c67-807a-ea97a3df9398\") " pod="openshift-multus/multus-bwpqp" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.537005 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/662dcbe7-7cf8-42d4-9a5e-345c32a49972-cni-binary-copy\") pod \"multus-additional-cni-plugins-mfp2k\" (UID: \"662dcbe7-7cf8-42d4-9a5e-345c32a49972\") " pod="openshift-multus/multus-additional-cni-plugins-mfp2k" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.537006 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/5b5233d6-d2b8-42dc-b881-18e439ecc442-ovnkube-config\") pod \"ovnkube-control-plane-749d76644c-4xdd8\" (UID: \"5b5233d6-d2b8-42dc-b881-18e439ecc442\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xdd8" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.535374 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-host-run-netns\") pod \"ovnkube-node-f7mp5\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.537059 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"hostroot\" (UniqueName: \"kubernetes.io/host-path/3b7580c2-5c23-4c67-807a-ea97a3df9398-hostroot\") pod \"multus-bwpqp\" (UID: \"3b7580c2-5c23-4c67-807a-ea97a3df9398\") " pod="openshift-multus/multus-bwpqp" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.537084 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-k8s-cni-cncf-io\" (UniqueName: \"kubernetes.io/host-path/3b7580c2-5c23-4c67-807a-ea97a3df9398-host-run-k8s-cni-cncf-io\") pod \"multus-bwpqp\" (UID: \"3b7580c2-5c23-4c67-807a-ea97a3df9398\") " pod="openshift-multus/multus-bwpqp" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.535966 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/3b7580c2-5c23-4c67-807a-ea97a3df9398-multus-daemon-config\") pod \"multus-bwpqp\" (UID: \"3b7580c2-5c23-4c67-807a-ea97a3df9398\") " pod="openshift-multus/multus-bwpqp" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.537124 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-etc-kube\" (UniqueName: \"kubernetes.io/host-path/37a5e44f-9a88-4405-be8a-b645485e7312-host-etc-kube\") pod \"network-operator-58b4c7f79c-55gtf\" (UID: \"37a5e44f-9a88-4405-be8a-b645485e7312\") " pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.537155 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cnibin\" (UniqueName: \"kubernetes.io/host-path/662dcbe7-7cf8-42d4-9a5e-345c32a49972-cnibin\") pod \"multus-additional-cni-plugins-mfp2k\" (UID: \"662dcbe7-7cf8-42d4-9a5e-345c32a49972\") " pod="openshift-multus/multus-additional-cni-plugins-mfp2k" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.537451 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-socket-dir-parent\" (UniqueName: \"kubernetes.io/host-path/3b7580c2-5c23-4c67-807a-ea97a3df9398-multus-socket-dir-parent\") pod \"multus-bwpqp\" (UID: \"3b7580c2-5c23-4c67-807a-ea97a3df9398\") " pod="openshift-multus/multus-bwpqp" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.537490 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tuning-conf-dir\" (UniqueName: \"kubernetes.io/host-path/662dcbe7-7cf8-42d4-9a5e-345c32a49972-tuning-conf-dir\") pod \"multus-additional-cni-plugins-mfp2k\" (UID: \"662dcbe7-7cf8-42d4-9a5e-345c32a49972\") " pod="openshift-multus/multus-additional-cni-plugins-mfp2k" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.537541 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-run-openvswitch\") pod \"ovnkube-node-f7mp5\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.536061 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-var-lib-openvswitch\") pod \"ovnkube-node-f7mp5\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.538018 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/662dcbe7-7cf8-42d4-9a5e-345c32a49972-os-release\") pod \"multus-additional-cni-plugins-mfp2k\" (UID: \"662dcbe7-7cf8-42d4-9a5e-345c32a49972\") " pod="openshift-multus/multus-additional-cni-plugins-mfp2k" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.538045 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-run-systemd\") pod \"ovnkube-node-f7mp5\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.539061 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/d0d6440c-4524-42ea-b18a-0ddc4662f672-serviceca\") pod \"node-ca-njc9c\" (UID: \"d0d6440c-4524-42ea-b18a-0ddc4662f672\") " pod="openshift-image-registry/node-ca-njc9c" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.538092 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/d75a4c96-2883-4a0b-bab2-0fab2b6c0b49-host-slash\") pod \"iptables-alerter-4ln5h\" (UID: \"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\") " pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.538062 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-host-run-ovn-kubernetes\") pod \"ovnkube-node-f7mp5\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.539985 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5b5233d6-d2b8-42dc-b881-18e439ecc442-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-4xdd8\" (UID: \"5b5233d6-d2b8-42dc-b881-18e439ecc442\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xdd8" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.540044 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-host-kubelet\") pod \"ovnkube-node-f7mp5\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.540081 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-run-ovn\") pod \"ovnkube-node-f7mp5\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.540112 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/3b7580c2-5c23-4c67-807a-ea97a3df9398-os-release\") pod \"multus-bwpqp\" (UID: \"3b7580c2-5c23-4c67-807a-ea97a3df9398\") " pod="openshift-multus/multus-bwpqp" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.540142 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/3b7580c2-5c23-4c67-807a-ea97a3df9398-host-var-lib-kubelet\") pod \"multus-bwpqp\" (UID: \"3b7580c2-5c23-4c67-807a-ea97a3df9398\") " pod="openshift-multus/multus-bwpqp" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.540172 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/3b7580c2-5c23-4c67-807a-ea97a3df9398-multus-conf-dir\") pod \"multus-bwpqp\" (UID: \"3b7580c2-5c23-4c67-807a-ea97a3df9398\") " pod="openshift-multus/multus-bwpqp" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.540201 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-etc-openvswitch\") pod \"ovnkube-node-f7mp5\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.540229 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-log-socket\") pod \"ovnkube-node-f7mp5\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.540258 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/15117613-9fad-48c7-98c4-a2d84502ded9-ovn-node-metrics-cert\") pod \"ovnkube-node-f7mp5\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.540288 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/15117613-9fad-48c7-98c4-a2d84502ded9-ovnkube-script-lib\") pod \"ovnkube-node-f7mp5\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.540343 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/be985ff1-de8a-431f-a230-67894ebe52d5-metrics-certs\") pod \"network-metrics-daemon-mqmtd\" (UID: \"be985ff1-de8a-431f-a230-67894ebe52d5\") " pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.540375 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5b5233d6-d2b8-42dc-b881-18e439ecc442-env-overrides\") pod \"ovnkube-control-plane-749d76644c-4xdd8\" (UID: \"5b5233d6-d2b8-42dc-b881-18e439ecc442\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xdd8" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.540405 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/3b7580c2-5c23-4c67-807a-ea97a3df9398-host-run-multus-certs\") pod \"multus-bwpqp\" (UID: \"3b7580c2-5c23-4c67-807a-ea97a3df9398\") " pod="openshift-multus/multus-bwpqp" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.540436 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s8lxm\" (UniqueName: \"kubernetes.io/projected/3b7580c2-5c23-4c67-807a-ea97a3df9398-kube-api-access-s8lxm\") pod \"multus-bwpqp\" (UID: \"3b7580c2-5c23-4c67-807a-ea97a3df9398\") " pod="openshift-multus/multus-bwpqp" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.540511 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/3b7580c2-5c23-4c67-807a-ea97a3df9398-cni-binary-copy\") pod \"multus-bwpqp\" (UID: \"3b7580c2-5c23-4c67-807a-ea97a3df9398\") " pod="openshift-multus/multus-bwpqp" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.540543 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3b7580c2-5c23-4c67-807a-ea97a3df9398-host-var-lib-cni-bin\") pod \"multus-bwpqp\" (UID: \"3b7580c2-5c23-4c67-807a-ea97a3df9398\") " pod="openshift-multus/multus-bwpqp" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.540580 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/662dcbe7-7cf8-42d4-9a5e-345c32a49972-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-mfp2k\" (UID: \"662dcbe7-7cf8-42d4-9a5e-345c32a49972\") " pod="openshift-multus/multus-additional-cni-plugins-mfp2k" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.540614 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b6psg\" (UniqueName: \"kubernetes.io/projected/662dcbe7-7cf8-42d4-9a5e-345c32a49972-kube-api-access-b6psg\") pod \"multus-additional-cni-plugins-mfp2k\" (UID: \"662dcbe7-7cf8-42d4-9a5e-345c32a49972\") " pod="openshift-multus/multus-additional-cni-plugins-mfp2k" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.540719 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.540742 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.540762 4948 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/8f668bae-612b-4b75-9490-919e737c6a3b-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.540783 4948 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.540801 4948 reconciler_common.go:293] "Volume detached for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-available-featuregates\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.540821 4948 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.540839 4948 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/6402fda4-df10-493c-b4e5-d0569419652d-images\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.540856 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-279lb\" (UniqueName: \"kubernetes.io/projected/7bb08738-c794-4ee8-9972-3a62ca171029-kube-api-access-279lb\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.540875 4948 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/496e6271-fb68-4057-954e-a0d97a4afa3f-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.540894 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.540911 4948 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.540930 4948 reconciler_common.go:293] "Volume detached for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/efdd0498-1daa-4136-9a4a-3b948c2293fc-webhook-certs\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.540950 4948 reconciler_common.go:293] "Volume detached for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-sysctl-allowlist\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.540969 4948 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.540987 4948 reconciler_common.go:293] "Volume detached for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/0b574797-001e-440a-8f4e-c0be86edad0f-mcc-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.541005 4948 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.541023 4948 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.541041 4948 reconciler_common.go:293] "Volume detached for volume \"multus-daemon-config\" (UniqueName: \"kubernetes.io/configmap/4bb40260-dbaa-4fb0-84df-5e680505d512-multus-daemon-config\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.541059 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jkwtn\" (UniqueName: \"kubernetes.io/projected/5b88f790-22fa-440e-b583-365168c0b23d-kube-api-access-jkwtn\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.541076 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rnphk\" (UniqueName: \"kubernetes.io/projected/bf126b07-da06-4140-9a57-dfd54fc6b486-kube-api-access-rnphk\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.541096 4948 reconciler_common.go:293] "Volume detached for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/c03ee662-fb2f-4fc4-a2c1-af487c19d254-stats-auth\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.541112 4948 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/01ab3dd5-8196-46d0-ad33-122e2ca51def-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.541130 4948 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/a31745f5-9847-4afe-82a5-3161cc66ca93-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.541147 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wxkg8\" (UniqueName: \"kubernetes.io/projected/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59-kube-api-access-wxkg8\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.541165 4948 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.541184 4948 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7539238d-5fe0-46ed-884e-1c3b566537ec-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.541202 4948 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/5441d097-087c-4d9a-baa8-b210afa90fc9-client-ca\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.541220 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-htfz6\" (UniqueName: \"kubernetes.io/projected/6ea678ab-3438-413e-bfe3-290ae7725660-kube-api-access-htfz6\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.541251 4948 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.541268 4948 reconciler_common.go:293] "Volume detached for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/6402fda4-df10-493c-b4e5-d0569419652d-machine-api-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.541286 4948 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.541327 4948 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/6ea678ab-3438-413e-bfe3-290ae7725660-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.541347 4948 reconciler_common.go:293] "Volume detached for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-webhook-cert\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.541376 4948 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/87cf06ed-a83f-41a7-828d-70653580a8cb-config-volume\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.541426 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5225d0e4-402f-4861-b410-819f433b1803-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.541500 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w7l8j\" (UniqueName: \"kubernetes.io/projected/01ab3dd5-8196-46d0-ad33-122e2ca51def-kube-api-access-w7l8j\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.541531 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6ccd8\" (UniqueName: \"kubernetes.io/projected/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-kube-api-access-6ccd8\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.541555 4948 reconciler_common.go:293] "Volume detached for volume \"certs\" (UniqueName: \"kubernetes.io/secret/5fe579f8-e8a6-4643-bce5-a661393c4dde-certs\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.541556 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/15117613-9fad-48c7-98c4-a2d84502ded9-ovnkube-script-lib\") pod \"ovnkube-node-f7mp5\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.541580 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dbsvg\" (UniqueName: \"kubernetes.io/projected/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9-kube-api-access-dbsvg\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.541597 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-run-ovn\") pod \"ovnkube-node-f7mp5\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.541633 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-host-kubelet\") pod \"ovnkube-node-f7mp5\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.541658 4948 reconciler_common.go:293] "Volume detached for volume \"images\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-images\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.541695 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"os-release\" (UniqueName: \"kubernetes.io/host-path/3b7580c2-5c23-4c67-807a-ea97a3df9398-os-release\") pod \"multus-bwpqp\" (UID: \"3b7580c2-5c23-4c67-807a-ea97a3df9398\") " pod="openshift-multus/multus-bwpqp" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.541713 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-etc-openvswitch\") pod \"ovnkube-node-f7mp5\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.541728 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-log-socket\") pod \"ovnkube-node-f7mp5\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:14 crc kubenswrapper[4948]: E0312 00:08:14.541787 4948 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.541802 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mg5zb\" (UniqueName: \"kubernetes.io/projected/6402fda4-df10-493c-b4e5-d0569419652d-kube-api-access-mg5zb\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: E0312 00:08:14.541840 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/be985ff1-de8a-431f-a230-67894ebe52d5-metrics-certs podName:be985ff1-de8a-431f-a230-67894ebe52d5 nodeName:}" failed. No retries permitted until 2026-03-12 00:08:15.041826378 +0000 UTC m=+94.497430106 (durationBeforeRetry 500ms). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/be985ff1-de8a-431f-a230-67894ebe52d5-metrics-certs") pod "network-metrics-daemon-mqmtd" (UID: "be985ff1-de8a-431f-a230-67894ebe52d5") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.541913 4948 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.541930 4948 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.541941 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fcqwp\" (UniqueName: \"kubernetes.io/projected/5fe579f8-e8a6-4643-bce5-a661393c4dde-kube-api-access-fcqwp\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.541951 4948 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.541962 4948 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/bf126b07-da06-4140-9a57-dfd54fc6b486-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.541972 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fqsjt\" (UniqueName: \"kubernetes.io/projected/efdd0498-1daa-4136-9a4a-3b948c2293fc-kube-api-access-fqsjt\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.541981 4948 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.541990 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-9xfj7\" (UniqueName: \"kubernetes.io/projected/5225d0e4-402f-4861-b410-819f433b1803-kube-api-access-9xfj7\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.542001 4948 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/0b78653f-4ff9-4508-8672-245ed9b561e3-service-ca\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.542011 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x2m85\" (UniqueName: \"kubernetes.io/projected/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d-kube-api-access-x2m85\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.542021 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mnrrd\" (UniqueName: \"kubernetes.io/projected/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d-kube-api-access-mnrrd\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.542032 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/01ab3dd5-8196-46d0-ad33-122e2ca51def-config\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.542040 4948 reconciler_common.go:293] "Volume detached for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-ca\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.542049 4948 reconciler_common.go:293] "Volume detached for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/a0128f3a-b052-44ed-a84e-c4c8aaf17c13-samples-operator-tls\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.542058 4948 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.542068 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-249nr\" (UniqueName: \"kubernetes.io/projected/b6312bbd-5731-4ea0-a20f-81d5a57df44a-kube-api-access-249nr\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.542078 4948 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.542087 4948 reconciler_common.go:293] "Volume detached for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-oauth-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.542096 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tk88c\" (UniqueName: \"kubernetes.io/projected/7539238d-5fe0-46ed-884e-1c3b566537ec-kube-api-access-tk88c\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.542105 4948 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-encryption-config\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.542115 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/1d611f23-29be-4491-8495-bee1670e935f-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.542125 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pj782\" (UniqueName: \"kubernetes.io/projected/b6cd30de-2eeb-49a2-ab40-9167f4560ff5-kube-api-access-pj782\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.542138 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jhbk2\" (UniqueName: \"kubernetes.io/projected/bd23aa5c-e532-4e53-bccf-e79f130c5ae8-kube-api-access-jhbk2\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.542150 4948 reconciler_common.go:293] "Volume detached for volume \"cert\" (UniqueName: \"kubernetes.io/secret/20b0d48f-5fd6-431c-a545-e3c800c7b866-cert\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.542159 4948 reconciler_common.go:293] "Volume detached for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/b6312bbd-5731-4ea0-a20f-81d5a57df44a-profile-collector-cert\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.542170 4948 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-registry-tls\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.542179 4948 reconciler_common.go:293] "Volume detached for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-service-ca\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.542189 4948 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/8f668bae-612b-4b75-9490-919e737c6a3b-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.542198 4948 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.542208 4948 reconciler_common.go:293] "Volume detached for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-apiservice-cert\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.542217 4948 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/6ea678ab-3438-413e-bfe3-290ae7725660-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.542226 4948 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/fda69060-fa79-4696-b1a6-7980f124bf7c-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.542234 4948 reconciler_common.go:293] "Volume detached for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-serving-ca\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.542244 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kfwg7\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-kube-api-access-kfwg7\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.542241 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/5b5233d6-d2b8-42dc-b881-18e439ecc442-env-overrides\") pod \"ovnkube-control-plane-749d76644c-4xdd8\" (UID: \"5b5233d6-d2b8-42dc-b881-18e439ecc442\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xdd8" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.540517 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"multus-conf-dir\" (UniqueName: \"kubernetes.io/host-path/3b7580c2-5c23-4c67-807a-ea97a3df9398-multus-conf-dir\") pod \"multus-bwpqp\" (UID: \"3b7580c2-5c23-4c67-807a-ea97a3df9398\") " pod="openshift-multus/multus-bwpqp" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.542254 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-xcgwh\" (UniqueName: \"kubernetes.io/projected/fda69060-fa79-4696-b1a6-7980f124bf7c-kube-api-access-xcgwh\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.542335 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-kubelet\" (UniqueName: \"kubernetes.io/host-path/3b7580c2-5c23-4c67-807a-ea97a3df9398-host-var-lib-kubelet\") pod \"multus-bwpqp\" (UID: \"3b7580c2-5c23-4c67-807a-ea97a3df9398\") " pod="openshift-multus/multus-bwpqp" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.542387 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-bin\" (UniqueName: \"kubernetes.io/host-path/3b7580c2-5c23-4c67-807a-ea97a3df9398-host-var-lib-cni-bin\") pod \"multus-bwpqp\" (UID: \"3b7580c2-5c23-4c67-807a-ea97a3df9398\") " pod="openshift-multus/multus-bwpqp" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.542418 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-multus-certs\" (UniqueName: \"kubernetes.io/host-path/3b7580c2-5c23-4c67-807a-ea97a3df9398-host-run-multus-certs\") pod \"multus-bwpqp\" (UID: \"3b7580c2-5c23-4c67-807a-ea97a3df9398\") " pod="openshift-multus/multus-bwpqp" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.542644 4948 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.542669 4948 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1386a44e-36a2-460c-96d0-0359d2b6f0f5-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.542688 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/57a731c4-ef35-47a8-b875-bfb08a7f8011-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.542709 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pcxfs\" (UniqueName: \"kubernetes.io/projected/9d4552c7-cd75-42dd-8880-30dd377c49a4-kube-api-access-pcxfs\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.542728 4948 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/1bf7eb37-55a3-4c65-b768-a94c82151e69-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.542747 4948 reconciler_common.go:293] "Volume detached for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-service-ca\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.542765 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7583ce53-e0fe-4a16-9e4d-50516596a136-config\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.542784 4948 reconciler_common.go:293] "Volume detached for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/43509403-f426-496e-be36-56cef71462f5-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.542802 4948 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7583ce53-e0fe-4a16-9e4d-50516596a136-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.542820 4948 reconciler_common.go:293] "Volume detached for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b-tmpfs\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.542838 4948 reconciler_common.go:293] "Volume detached for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/fda69060-fa79-4696-b1a6-7980f124bf7c-mcd-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.542856 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1386a44e-36a2-460c-96d0-0359d2b6f0f5-config\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.542877 4948 reconciler_common.go:293] "Volume detached for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/87cf06ed-a83f-41a7-828d-70653580a8cb-metrics-tls\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.542858 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-sysctl-allowlist\" (UniqueName: \"kubernetes.io/configmap/662dcbe7-7cf8-42d4-9a5e-345c32a49972-cni-sysctl-allowlist\") pod \"multus-additional-cni-plugins-mfp2k\" (UID: \"662dcbe7-7cf8-42d4-9a5e-345c32a49972\") " pod="openshift-multus/multus-additional-cni-plugins-mfp2k" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.542896 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1bf7eb37-55a3-4c65-b768-a94c82151e69-config\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.542915 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-config\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.542934 4948 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.542955 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w9rds\" (UniqueName: \"kubernetes.io/projected/20b0d48f-5fd6-431c-a545-e3c800c7b866-kube-api-access-w9rds\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.542978 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zgdk5\" (UniqueName: \"kubernetes.io/projected/31d8b7a1-420e-4252-a5b7-eebe8a111292-kube-api-access-zgdk5\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.542997 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ngvvp\" (UniqueName: \"kubernetes.io/projected/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-kube-api-access-ngvvp\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.543015 4948 reconciler_common.go:293] "Volume detached for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-encryption-config\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.542863 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/3b7580c2-5c23-4c67-807a-ea97a3df9398-cni-binary-copy\") pod \"multus-bwpqp\" (UID: \"3b7580c2-5c23-4c67-807a-ea97a3df9398\") " pod="openshift-multus/multus-bwpqp" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.543032 4948 reconciler_common.go:293] "Volume detached for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/3ab1a177-2de0-46d9-b765-d0d0649bb42e-package-server-manager-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.543052 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-bf2bz\" (UniqueName: \"kubernetes.io/projected/1d611f23-29be-4491-8495-bee1670e935f-kube-api-access-bf2bz\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.543072 4948 reconciler_common.go:293] "Volume detached for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/6509e943-70c6-444c-bc41-48a544e36fbd-service-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.544026 4948 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/31d8b7a1-420e-4252-a5b7-eebe8a111292-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.544090 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8tdtz\" (UniqueName: \"kubernetes.io/projected/09efc573-dbb6-4249-bd59-9b87aba8dd28-kube-api-access-8tdtz\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.544107 4948 reconciler_common.go:293] "Volume detached for volume \"cni-binary-copy\" (UniqueName: \"kubernetes.io/configmap/7bb08738-c794-4ee8-9972-3a62ca171029-cni-binary-copy\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.544125 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-7c4vf\" (UniqueName: \"kubernetes.io/projected/22c825df-677d-4ca6-82db-3454ed06e783-kube-api-access-7c4vf\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.544145 4948 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/09efc573-dbb6-4249-bd59-9b87aba8dd28-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.544162 4948 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/31d8b7a1-420e-4252-a5b7-eebe8a111292-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.544180 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d4lsv\" (UniqueName: \"kubernetes.io/projected/25e176fe-21b4-4974-b1ed-c8b94f112a7f-kube-api-access-d4lsv\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.544197 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2d4wz\" (UniqueName: \"kubernetes.io/projected/5441d097-087c-4d9a-baa8-b210afa90fc9-kube-api-access-2d4wz\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.544214 4948 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/49c341d1-5089-4bc2-86a0-a5e165cfcc6b-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.544235 4948 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/8f668bae-612b-4b75-9490-919e737c6a3b-registry-certificates\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.544251 4948 reconciler_common.go:293] "Volume detached for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/09ae3b1a-e8e7-4524-b54b-61eab6f9239a-etcd-client\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.544268 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qg5z5\" (UniqueName: \"kubernetes.io/projected/43509403-f426-496e-be36-56cef71462f5-kube-api-access-qg5z5\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.544286 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d6qdx\" (UniqueName: \"kubernetes.io/projected/87cf06ed-a83f-41a7-828d-70653580a8cb-kube-api-access-d6qdx\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.544322 4948 reconciler_common.go:293] "Volume detached for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/22c825df-677d-4ca6-82db-3454ed06e783-auth-proxy-config\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.544340 4948 reconciler_common.go:293] "Volume detached for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/22c825df-677d-4ca6-82db-3454ed06e783-machine-approver-tls\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.544357 4948 reconciler_common.go:293] "Volume detached for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/0b574797-001e-440a-8f4e-c0be86edad0f-proxy-tls\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.544373 4948 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/8f668bae-612b-4b75-9490-919e737c6a3b-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.543421 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcd-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/7e483c2b-08f0-4e92-8e4a-b7281f30af3e-mcd-auth-proxy-config\") pod \"machine-config-daemon-m4xwc\" (UID: \"7e483c2b-08f0-4e92-8e4a-b7281f30af3e\") " pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.545921 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/15117613-9fad-48c7-98c4-a2d84502ded9-env-overrides\") pod \"ovnkube-node-f7mp5\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.546000 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwpqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b7580c2-5c23-4c67-807a-ea97a3df9398\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8lxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwpqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.546113 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"system-cni-dir\" (UniqueName: \"kubernetes.io/host-path/662dcbe7-7cf8-42d4-9a5e-345c32a49972-system-cni-dir\") pod \"multus-additional-cni-plugins-mfp2k\" (UID: \"662dcbe7-7cf8-42d4-9a5e-345c32a49972\") " pod="openshift-multus/multus-additional-cni-plugins-mfp2k" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.549318 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dvwhv\" (UniqueName: \"kubernetes.io/projected/15117613-9fad-48c7-98c4-a2d84502ded9-kube-api-access-dvwhv\") pod \"ovnkube-node-f7mp5\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.549494 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/15117613-9fad-48c7-98c4-a2d84502ded9-ovn-node-metrics-cert\") pod \"ovnkube-node-f7mp5\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.550126 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9xf4z\" (UniqueName: \"kubernetes.io/projected/b9f9c7f1-6593-4d92-854c-e6c9964a0435-kube-api-access-9xf4z\") pod \"node-resolver-9s5fr\" (UID: \"b9f9c7f1-6593-4d92-854c-e6c9964a0435\") " pod="openshift-dns/node-resolver-9s5fr" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.556117 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-control-plane-metrics-cert\" (UniqueName: \"kubernetes.io/secret/5b5233d6-d2b8-42dc-b881-18e439ecc442-ovn-control-plane-metrics-cert\") pod \"ovnkube-control-plane-749d76644c-4xdd8\" (UID: \"5b5233d6-d2b8-42dc-b881-18e439ecc442\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xdd8" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.556382 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-njc9c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0d6440c-4524-42ea-b18a-0ddc4662f672\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7wn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-njc9c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.559779 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k7wn2\" (UniqueName: \"kubernetes.io/projected/d0d6440c-4524-42ea-b18a-0ddc4662f672-kube-api-access-k7wn2\") pod \"node-ca-njc9c\" (UID: \"d0d6440c-4524-42ea-b18a-0ddc4662f672\") " pod="openshift-image-registry/node-ca-njc9c" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.560633 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/7e483c2b-08f0-4e92-8e4a-b7281f30af3e-proxy-tls\") pod \"machine-config-daemon-m4xwc\" (UID: \"7e483c2b-08f0-4e92-8e4a-b7281f30af3e\") " pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.561141 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hjpwd\" (UniqueName: \"kubernetes.io/projected/5b5233d6-d2b8-42dc-b881-18e439ecc442-kube-api-access-hjpwd\") pod \"ovnkube-control-plane-749d76644c-4xdd8\" (UID: \"5b5233d6-d2b8-42dc-b881-18e439ecc442\") " pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xdd8" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.561204 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gd9jv\" (UniqueName: \"kubernetes.io/projected/7e483c2b-08f0-4e92-8e4a-b7281f30af3e-kube-api-access-gd9jv\") pod \"machine-config-daemon-m4xwc\" (UID: \"7e483c2b-08f0-4e92-8e4a-b7281f30af3e\") " pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.562904 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b6psg\" (UniqueName: \"kubernetes.io/projected/662dcbe7-7cf8-42d4-9a5e-345c32a49972-kube-api-access-b6psg\") pod \"multus-additional-cni-plugins-mfp2k\" (UID: \"662dcbe7-7cf8-42d4-9a5e-345c32a49972\") " pod="openshift-multus/multus-additional-cni-plugins-mfp2k" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.563329 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hmsz5\" (UniqueName: \"kubernetes.io/projected/be985ff1-de8a-431f-a230-67894ebe52d5-kube-api-access-hmsz5\") pod \"network-metrics-daemon-mqmtd\" (UID: \"be985ff1-de8a-431f-a230-67894ebe52d5\") " pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.565380 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xdd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b5233d6-d2b8-42dc-b881-18e439ecc442\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjpwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjpwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4xdd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.572096 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9s5fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f9c7f1-6593-4d92-854c-e6c9964a0435\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xf4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9s5fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": dial tcp 127.0.0.1:9743: connect: connection refused" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.574468 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s8lxm\" (UniqueName: \"kubernetes.io/projected/3b7580c2-5c23-4c67-807a-ea97a3df9398-kube-api-access-s8lxm\") pod \"multus-bwpqp\" (UID: \"3b7580c2-5c23-4c67-807a-ea97a3df9398\") " pod="openshift-multus/multus-bwpqp" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.630721 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.630771 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.630780 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.630792 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.630801 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:14Z","lastTransitionTime":"2026-03-12T00:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.642942 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-node-identity/network-node-identity-vrzqb" Mar 12 00:08:14 crc kubenswrapper[4948]: W0312 00:08:14.663587 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podef543e1b_8068_4ea3_b32a_61027b32e95d.slice/crio-21b895e05ab3dd0ccb1d807401c64c1275b121896344e54b8ba5703688d08e10 WatchSource:0}: Error finding container 21b895e05ab3dd0ccb1d807401c64c1275b121896344e54b8ba5703688d08e10: Status 404 returned error can't find the container with id 21b895e05ab3dd0ccb1d807401c64c1275b121896344e54b8ba5703688d08e10 Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.679213 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/iptables-alerter-4ln5h" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.694871 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" Mar 12 00:08:14 crc kubenswrapper[4948]: W0312 00:08:14.696718 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd75a4c96_2883_4a0b_bab2_0fab2b6c0b49.slice/crio-97dd4b7b8658996905b79ebbbb068431bee17f977a0b695693d8ba818accd3e5 WatchSource:0}: Error finding container 97dd4b7b8658996905b79ebbbb068431bee17f977a0b695693d8ba818accd3e5: Status 404 returned error can't find the container with id 97dd4b7b8658996905b79ebbbb068431bee17f977a0b695693d8ba818accd3e5 Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.709526 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-bwpqp" Mar 12 00:08:14 crc kubenswrapper[4948]: W0312 00:08:14.717190 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod37a5e44f_9a88_4405_be8a_b645485e7312.slice/crio-4c6539e3ae80b8321a0940db3a60894508ad25d8af5227099137776dae71e34e WatchSource:0}: Error finding container 4c6539e3ae80b8321a0940db3a60894508ad25d8af5227099137776dae71e34e: Status 404 returned error can't find the container with id 4c6539e3ae80b8321a0940db3a60894508ad25d8af5227099137776dae71e34e Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.724141 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/node-resolver-9s5fr" Mar 12 00:08:14 crc kubenswrapper[4948]: W0312 00:08:14.726475 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3b7580c2_5c23_4c67_807a_ea97a3df9398.slice/crio-9acc261fa502de9cfe0b3c954de34929d3e6d671d8e4eaff8c891ab737ab5a4b WatchSource:0}: Error finding container 9acc261fa502de9cfe0b3c954de34929d3e6d671d8e4eaff8c891ab737ab5a4b: Status 404 returned error can't find the container with id 9acc261fa502de9cfe0b3c954de34929d3e6d671d8e4eaff8c891ab737ab5a4b Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.733186 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.733223 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.733231 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.733244 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.733254 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:14Z","lastTransitionTime":"2026-03-12T00:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:14 crc kubenswrapper[4948]: W0312 00:08:14.751637 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podb9f9c7f1_6593_4d92_854c_e6c9964a0435.slice/crio-3f23525705787a472386c3f098f6e35896103fbc5474efcd65546f8447b89038 WatchSource:0}: Error finding container 3f23525705787a472386c3f098f6e35896103fbc5474efcd65546f8447b89038: Status 404 returned error can't find the container with id 3f23525705787a472386c3f098f6e35896103fbc5474efcd65546f8447b89038 Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.760463 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xdd8" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.767936 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.772976 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-additional-cni-plugins-mfp2k" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.779207 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/node-ca-njc9c" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.783838 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" Mar 12 00:08:14 crc kubenswrapper[4948]: W0312 00:08:14.806616 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod7e483c2b_08f0_4e92_8e4a_b7281f30af3e.slice/crio-82bbea1aab6ebe034fed5428a33876c0d074c5676beb3be8b6a4b590aa0d7be6 WatchSource:0}: Error finding container 82bbea1aab6ebe034fed5428a33876c0d074c5676beb3be8b6a4b590aa0d7be6: Status 404 returned error can't find the container with id 82bbea1aab6ebe034fed5428a33876c0d074c5676beb3be8b6a4b590aa0d7be6 Mar 12 00:08:14 crc kubenswrapper[4948]: W0312 00:08:14.822766 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod662dcbe7_7cf8_42d4_9a5e_345c32a49972.slice/crio-22d2ae7181e1db8696c6685f19a54ddc9f9d8bdb3e460920d24c957f5d0353e7 WatchSource:0}: Error finding container 22d2ae7181e1db8696c6685f19a54ddc9f9d8bdb3e460920d24c957f5d0353e7: Status 404 returned error can't find the container with id 22d2ae7181e1db8696c6685f19a54ddc9f9d8bdb3e460920d24c957f5d0353e7 Mar 12 00:08:14 crc kubenswrapper[4948]: W0312 00:08:14.825691 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod5b5233d6_d2b8_42dc_b881_18e439ecc442.slice/crio-b398f92a25947535f904350a049601c7bd8e14c799616b1f1a7d5bcd19c60cbb WatchSource:0}: Error finding container b398f92a25947535f904350a049601c7bd8e14c799616b1f1a7d5bcd19c60cbb: Status 404 returned error can't find the container with id b398f92a25947535f904350a049601c7bd8e14c799616b1f1a7d5bcd19c60cbb Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.837524 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.837611 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.837628 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.837682 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.837700 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:14Z","lastTransitionTime":"2026-03-12T00:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:14 crc kubenswrapper[4948]: W0312 00:08:14.837771 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podd0d6440c_4524_42ea_b18a_0ddc4662f672.slice/crio-e1d2ee256162a19281dd9d8a7eee328f3c1b78581ef324109ac878b2f432641b WatchSource:0}: Error finding container e1d2ee256162a19281dd9d8a7eee328f3c1b78581ef324109ac878b2f432641b: Status 404 returned error can't find the container with id e1d2ee256162a19281dd9d8a7eee328f3c1b78581ef324109ac878b2f432641b Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.941095 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.941125 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.941133 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.941145 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.941154 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:14Z","lastTransitionTime":"2026-03-12T00:08:14Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.947882 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.948017 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.948077 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.948114 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:08:14 crc kubenswrapper[4948]: I0312 00:08:14.948139 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:08:14 crc kubenswrapper[4948]: E0312 00:08:14.948349 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 12 00:08:14 crc kubenswrapper[4948]: E0312 00:08:14.948377 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 12 00:08:14 crc kubenswrapper[4948]: E0312 00:08:14.948394 4948 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 00:08:14 crc kubenswrapper[4948]: E0312 00:08:14.948468 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-12 00:08:15.948452106 +0000 UTC m=+95.404055854 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 00:08:14 crc kubenswrapper[4948]: E0312 00:08:14.948898 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 00:08:15.948883407 +0000 UTC m=+95.404487165 (durationBeforeRetry 1s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:08:14 crc kubenswrapper[4948]: E0312 00:08:14.948967 4948 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 12 00:08:14 crc kubenswrapper[4948]: E0312 00:08:14.949000 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-12 00:08:15.948991142 +0000 UTC m=+95.404594900 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 12 00:08:14 crc kubenswrapper[4948]: E0312 00:08:14.949061 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 12 00:08:14 crc kubenswrapper[4948]: E0312 00:08:14.949081 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 12 00:08:14 crc kubenswrapper[4948]: E0312 00:08:14.949094 4948 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 00:08:14 crc kubenswrapper[4948]: E0312 00:08:14.949045 4948 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 12 00:08:14 crc kubenswrapper[4948]: E0312 00:08:14.949124 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-12 00:08:15.949114727 +0000 UTC m=+95.404718485 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 00:08:14 crc kubenswrapper[4948]: E0312 00:08:14.949708 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-12 00:08:15.949682135 +0000 UTC m=+95.405285883 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.042960 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.042994 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.043003 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.043016 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.043026 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:15Z","lastTransitionTime":"2026-03-12T00:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.048968 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/be985ff1-de8a-431f-a230-67894ebe52d5-metrics-certs\") pod \"network-metrics-daemon-mqmtd\" (UID: \"be985ff1-de8a-431f-a230-67894ebe52d5\") " pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:08:15 crc kubenswrapper[4948]: E0312 00:08:15.049084 4948 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 12 00:08:15 crc kubenswrapper[4948]: E0312 00:08:15.049126 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/be985ff1-de8a-431f-a230-67894ebe52d5-metrics-certs podName:be985ff1-de8a-431f-a230-67894ebe52d5 nodeName:}" failed. No retries permitted until 2026-03-12 00:08:16.049114764 +0000 UTC m=+95.504718502 (durationBeforeRetry 1s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/be985ff1-de8a-431f-a230-67894ebe52d5-metrics-certs") pod "network-metrics-daemon-mqmtd" (UID: "be985ff1-de8a-431f-a230-67894ebe52d5") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.147849 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.147913 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.147926 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.147944 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.147978 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:15Z","lastTransitionTime":"2026-03-12T00:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.250673 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.250713 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.250726 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.250751 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.250763 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:15Z","lastTransitionTime":"2026-03-12T00:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.315763 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="01ab3dd5-8196-46d0-ad33-122e2ca51def" path="/var/lib/kubelet/pods/01ab3dd5-8196-46d0-ad33-122e2ca51def/volumes" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.316491 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09ae3b1a-e8e7-4524-b54b-61eab6f9239a" path="/var/lib/kubelet/pods/09ae3b1a-e8e7-4524-b54b-61eab6f9239a/volumes" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.317928 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="09efc573-dbb6-4249-bd59-9b87aba8dd28" path="/var/lib/kubelet/pods/09efc573-dbb6-4249-bd59-9b87aba8dd28/volumes" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.318681 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b574797-001e-440a-8f4e-c0be86edad0f" path="/var/lib/kubelet/pods/0b574797-001e-440a-8f4e-c0be86edad0f/volumes" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.319787 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="0b78653f-4ff9-4508-8672-245ed9b561e3" path="/var/lib/kubelet/pods/0b78653f-4ff9-4508-8672-245ed9b561e3/volumes" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.320465 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1386a44e-36a2-460c-96d0-0359d2b6f0f5" path="/var/lib/kubelet/pods/1386a44e-36a2-460c-96d0-0359d2b6f0f5/volumes" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.321223 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1bf7eb37-55a3-4c65-b768-a94c82151e69" path="/var/lib/kubelet/pods/1bf7eb37-55a3-4c65-b768-a94c82151e69/volumes" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.322392 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="1d611f23-29be-4491-8495-bee1670e935f" path="/var/lib/kubelet/pods/1d611f23-29be-4491-8495-bee1670e935f/volumes" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.323214 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="20b0d48f-5fd6-431c-a545-e3c800c7b866" path="/var/lib/kubelet/pods/20b0d48f-5fd6-431c-a545-e3c800c7b866/volumes" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.324318 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c" path="/var/lib/kubelet/pods/210d8245-ebfc-4e3b-ac4a-e21ce76f9a7c/volumes" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.324926 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="22c825df-677d-4ca6-82db-3454ed06e783" path="/var/lib/kubelet/pods/22c825df-677d-4ca6-82db-3454ed06e783/volumes" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.326214 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25e176fe-21b4-4974-b1ed-c8b94f112a7f" path="/var/lib/kubelet/pods/25e176fe-21b4-4974-b1ed-c8b94f112a7f/volumes" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.326833 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="308be0ea-9f5f-4b29-aeb1-5abd31a0b17b" path="/var/lib/kubelet/pods/308be0ea-9f5f-4b29-aeb1-5abd31a0b17b/volumes" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.327398 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="31d8b7a1-420e-4252-a5b7-eebe8a111292" path="/var/lib/kubelet/pods/31d8b7a1-420e-4252-a5b7-eebe8a111292/volumes" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.328550 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3ab1a177-2de0-46d9-b765-d0d0649bb42e" path="/var/lib/kubelet/pods/3ab1a177-2de0-46d9-b765-d0d0649bb42e/volumes" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.329196 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3cb93b32-e0ae-4377-b9c8-fdb9842c6d59" path="/var/lib/kubelet/pods/3cb93b32-e0ae-4377-b9c8-fdb9842c6d59/volumes" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.330356 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="43509403-f426-496e-be36-56cef71462f5" path="/var/lib/kubelet/pods/43509403-f426-496e-be36-56cef71462f5/volumes" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.330866 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44663579-783b-4372-86d6-acf235a62d72" path="/var/lib/kubelet/pods/44663579-783b-4372-86d6-acf235a62d72/volumes" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.331778 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="496e6271-fb68-4057-954e-a0d97a4afa3f" path="/var/lib/kubelet/pods/496e6271-fb68-4057-954e-a0d97a4afa3f/volumes" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.332857 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49c341d1-5089-4bc2-86a0-a5e165cfcc6b" path="/var/lib/kubelet/pods/49c341d1-5089-4bc2-86a0-a5e165cfcc6b/volumes" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.333463 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="49ef4625-1d3a-4a9f-b595-c2433d32326d" path="/var/lib/kubelet/pods/49ef4625-1d3a-4a9f-b595-c2433d32326d/volumes" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.334616 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4bb40260-dbaa-4fb0-84df-5e680505d512" path="/var/lib/kubelet/pods/4bb40260-dbaa-4fb0-84df-5e680505d512/volumes" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.335179 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5225d0e4-402f-4861-b410-819f433b1803" path="/var/lib/kubelet/pods/5225d0e4-402f-4861-b410-819f433b1803/volumes" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.336279 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5441d097-087c-4d9a-baa8-b210afa90fc9" path="/var/lib/kubelet/pods/5441d097-087c-4d9a-baa8-b210afa90fc9/volumes" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.336946 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="57a731c4-ef35-47a8-b875-bfb08a7f8011" path="/var/lib/kubelet/pods/57a731c4-ef35-47a8-b875-bfb08a7f8011/volumes" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.337712 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5b88f790-22fa-440e-b583-365168c0b23d" path="/var/lib/kubelet/pods/5b88f790-22fa-440e-b583-365168c0b23d/volumes" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.338791 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5fe579f8-e8a6-4643-bce5-a661393c4dde" path="/var/lib/kubelet/pods/5fe579f8-e8a6-4643-bce5-a661393c4dde/volumes" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.339405 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6402fda4-df10-493c-b4e5-d0569419652d" path="/var/lib/kubelet/pods/6402fda4-df10-493c-b4e5-d0569419652d/volumes" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.340510 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6509e943-70c6-444c-bc41-48a544e36fbd" path="/var/lib/kubelet/pods/6509e943-70c6-444c-bc41-48a544e36fbd/volumes" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.341174 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6731426b-95fe-49ff-bb5f-40441049fde2" path="/var/lib/kubelet/pods/6731426b-95fe-49ff-bb5f-40441049fde2/volumes" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.342086 4948 kubelet_volumes.go:152] "Cleaned up orphaned volume subpath from pod" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volume-subpaths/run-systemd/ovnkube-controller/6" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.342283 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ea678ab-3438-413e-bfe3-290ae7725660" path="/var/lib/kubelet/pods/6ea678ab-3438-413e-bfe3-290ae7725660/volumes" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.344112 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7539238d-5fe0-46ed-884e-1c3b566537ec" path="/var/lib/kubelet/pods/7539238d-5fe0-46ed-884e-1c3b566537ec/volumes" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.345196 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7583ce53-e0fe-4a16-9e4d-50516596a136" path="/var/lib/kubelet/pods/7583ce53-e0fe-4a16-9e4d-50516596a136/volumes" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.345716 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7bb08738-c794-4ee8-9972-3a62ca171029" path="/var/lib/kubelet/pods/7bb08738-c794-4ee8-9972-3a62ca171029/volumes" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.347206 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="87cf06ed-a83f-41a7-828d-70653580a8cb" path="/var/lib/kubelet/pods/87cf06ed-a83f-41a7-828d-70653580a8cb/volumes" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.348120 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8cea82b4-6893-4ddc-af9f-1bb5ae425c5b" path="/var/lib/kubelet/pods/8cea82b4-6893-4ddc-af9f-1bb5ae425c5b/volumes" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.349074 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="925f1c65-6136-48ba-85aa-3a3b50560753" path="/var/lib/kubelet/pods/925f1c65-6136-48ba-85aa-3a3b50560753/volumes" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.349767 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="96b93a3a-6083-4aea-8eab-fe1aa8245ad9" path="/var/lib/kubelet/pods/96b93a3a-6083-4aea-8eab-fe1aa8245ad9/volumes" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.351187 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9d4552c7-cd75-42dd-8880-30dd377c49a4" path="/var/lib/kubelet/pods/9d4552c7-cd75-42dd-8880-30dd377c49a4/volumes" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.352157 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a0128f3a-b052-44ed-a84e-c4c8aaf17c13" path="/var/lib/kubelet/pods/a0128f3a-b052-44ed-a84e-c4c8aaf17c13/volumes" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.353085 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="a31745f5-9847-4afe-82a5-3161cc66ca93" path="/var/lib/kubelet/pods/a31745f5-9847-4afe-82a5-3161cc66ca93/volumes" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.353707 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.353748 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.353763 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.353787 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.353804 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:15Z","lastTransitionTime":"2026-03-12T00:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.353966 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b11524ee-3fca-4b1b-9cdf-6da289fdbc7d" path="/var/lib/kubelet/pods/b11524ee-3fca-4b1b-9cdf-6da289fdbc7d/volumes" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.355854 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6312bbd-5731-4ea0-a20f-81d5a57df44a" path="/var/lib/kubelet/pods/b6312bbd-5731-4ea0-a20f-81d5a57df44a/volumes" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.356509 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="b6cd30de-2eeb-49a2-ab40-9167f4560ff5" path="/var/lib/kubelet/pods/b6cd30de-2eeb-49a2-ab40-9167f4560ff5/volumes" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.357658 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bc5039c0-ea34-426b-a2b7-fbbc87b49a6d" path="/var/lib/kubelet/pods/bc5039c0-ea34-426b-a2b7-fbbc87b49a6d/volumes" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.358397 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd23aa5c-e532-4e53-bccf-e79f130c5ae8" path="/var/lib/kubelet/pods/bd23aa5c-e532-4e53-bccf-e79f130c5ae8/volumes" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.359774 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bf126b07-da06-4140-9a57-dfd54fc6b486" path="/var/lib/kubelet/pods/bf126b07-da06-4140-9a57-dfd54fc6b486/volumes" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.360547 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="c03ee662-fb2f-4fc4-a2c1-af487c19d254" path="/var/lib/kubelet/pods/c03ee662-fb2f-4fc4-a2c1-af487c19d254/volumes" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.361660 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d" path="/var/lib/kubelet/pods/cd70aa09-68dd-4d64-bd6f-156fe6d1dc6d/volumes" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.362255 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e7e6199b-1264-4501-8953-767f51328d08" path="/var/lib/kubelet/pods/e7e6199b-1264-4501-8953-767f51328d08/volumes" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.362916 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="efdd0498-1daa-4136-9a4a-3b948c2293fc" path="/var/lib/kubelet/pods/efdd0498-1daa-4136-9a4a-3b948c2293fc/volumes" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.364120 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f88749ec-7931-4ee7-b3fc-1ec5e11f92e9" path="/var/lib/kubelet/pods/f88749ec-7931-4ee7-b3fc-1ec5e11f92e9/volumes" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.364727 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="fda69060-fa79-4696-b1a6-7980f124bf7c" path="/var/lib/kubelet/pods/fda69060-fa79-4696-b1a6-7980f124bf7c/volumes" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.456365 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.456464 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.456482 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.456538 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.456559 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:15Z","lastTransitionTime":"2026-03-12T00:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.506954 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"a5a9ae9441886817563a5fc793bff7c9b8d77d406ceb1d1fb1eebd4307a1c64b"} Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.506996 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" event={"ID":"37a5e44f-9a88-4405-be8a-b645485e7312","Type":"ContainerStarted","Data":"4c6539e3ae80b8321a0940db3a60894508ad25d8af5227099137776dae71e34e"} Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.511377 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"97dd4b7b8658996905b79ebbbb068431bee17f977a0b695693d8ba818accd3e5"} Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.515679 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" event={"ID":"7e483c2b-08f0-4e92-8e4a-b7281f30af3e","Type":"ContainerStarted","Data":"5d6839c99ab958ad1f40636c70e29bc285292fdbee1e910cdf7f8ab43ca7e8db"} Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.515742 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" event={"ID":"7e483c2b-08f0-4e92-8e4a-b7281f30af3e","Type":"ContainerStarted","Data":"82de8f2b633f1b6c295149d077bfd2b308f5acaa429acf411ed16f456536d0ba"} Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.515801 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" event={"ID":"7e483c2b-08f0-4e92-8e4a-b7281f30af3e","Type":"ContainerStarted","Data":"82bbea1aab6ebe034fed5428a33876c0d074c5676beb3be8b6a4b590aa0d7be6"} Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.520691 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-bwpqp" event={"ID":"3b7580c2-5c23-4c67-807a-ea97a3df9398","Type":"ContainerStarted","Data":"9d07c603619f858177d4b007e785208853abcfe2dc6c1af435309030fa409010"} Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.520746 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-bwpqp" event={"ID":"3b7580c2-5c23-4c67-807a-ea97a3df9398","Type":"ContainerStarted","Data":"9acc261fa502de9cfe0b3c954de34929d3e6d671d8e4eaff8c891ab737ab5a4b"} Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.526218 4948 generic.go:334] "Generic (PLEG): container finished" podID="15117613-9fad-48c7-98c4-a2d84502ded9" containerID="c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a" exitCode=0 Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.526269 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" event={"ID":"15117613-9fad-48c7-98c4-a2d84502ded9","Type":"ContainerDied","Data":"c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a"} Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.526349 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" event={"ID":"15117613-9fad-48c7-98c4-a2d84502ded9","Type":"ContainerStarted","Data":"2b5e9ab6e8d41452dcec8681682d1d5e7bd916598cb0e81980f0fc3ac545e902"} Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.526455 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9s5fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f9c7f1-6593-4d92-854c-e6c9964a0435\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [dns-node-resolver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xf4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9s5fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:15Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.528194 4948 generic.go:334] "Generic (PLEG): container finished" podID="662dcbe7-7cf8-42d4-9a5e-345c32a49972" containerID="c4a57d476b5b00110168878bd1c19c32ef122525c0259ac0a70de31c8220996d" exitCode=0 Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.528271 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-mfp2k" event={"ID":"662dcbe7-7cf8-42d4-9a5e-345c32a49972","Type":"ContainerDied","Data":"c4a57d476b5b00110168878bd1c19c32ef122525c0259ac0a70de31c8220996d"} Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.528290 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-mfp2k" event={"ID":"662dcbe7-7cf8-42d4-9a5e-345c32a49972","Type":"ContainerStarted","Data":"22d2ae7181e1db8696c6685f19a54ddc9f9d8bdb3e460920d24c957f5d0353e7"} Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.533396 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-9s5fr" event={"ID":"b9f9c7f1-6593-4d92-854c-e6c9964a0435","Type":"ContainerStarted","Data":"51b8833cbaf3809463e18bfebc9d5bada5fb26823831cc742a3a27177662f60e"} Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.533438 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/node-resolver-9s5fr" event={"ID":"b9f9c7f1-6593-4d92-854c-e6c9964a0435","Type":"ContainerStarted","Data":"3f23525705787a472386c3f098f6e35896103fbc5474efcd65546f8447b89038"} Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.537205 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xdd8" event={"ID":"5b5233d6-d2b8-42dc-b881-18e439ecc442","Type":"ContainerStarted","Data":"cb23ff4bc484524e306561c0c574e2b2eee39fc80efcd9f44f1924fe42df1a03"} Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.537263 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xdd8" event={"ID":"5b5233d6-d2b8-42dc-b881-18e439ecc442","Type":"ContainerStarted","Data":"f0ea55d993c72b6d50a4829ecfff4adae7d0f625a90e40545c9e63201cdda29c"} Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.537283 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xdd8" event={"ID":"5b5233d6-d2b8-42dc-b881-18e439ecc442","Type":"ContainerStarted","Data":"b398f92a25947535f904350a049601c7bd8e14c799616b1f1a7d5bcd19c60cbb"} Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.540704 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"1f5139686878eb07d3abced92d0ccfe05be8907580ed75233b72fc2ad3415a5a"} Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.540746 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"15cbe3fa75b69a7176803295e16a314a8bc73f5797c874124d54e76d47fa1064"} Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.540759 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" event={"ID":"ef543e1b-8068-4ea3-b32a-61027b32e95d","Type":"ContainerStarted","Data":"21b895e05ab3dd0ccb1d807401c64c1275b121896344e54b8ba5703688d08e10"} Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.541316 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xdd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b5233d6-d2b8-42dc-b881-18e439ecc442\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-rbac-proxy ovnkube-cluster-manager]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjpwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjpwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4xdd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:15Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.543597 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-njc9c" event={"ID":"d0d6440c-4524-42ea-b18a-0ddc4662f672","Type":"ContainerStarted","Data":"6f4d0d9599ad62aaba35ba1414a806f3a163e29a3c34fd893ef5b14eae003591"} Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.544157 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/node-ca-njc9c" event={"ID":"d0d6440c-4524-42ea-b18a-0ddc4662f672","Type":"ContainerStarted","Data":"e1d2ee256162a19281dd9d8a7eee328f3c1b78581ef324109ac878b2f432641b"} Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.558042 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:15Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.565037 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.565071 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.565082 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.565097 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.565108 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:15Z","lastTransitionTime":"2026-03-12T00:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.571814 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:15Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.584936 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mqmtd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be985ff1-de8a-431f-a230-67894ebe52d5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmsz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmsz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mqmtd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:15Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.606803 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15117613-9fad-48c7-98c4-a2d84502ded9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [kubecfg-setup]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7mp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:15Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.622988 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:15Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.642913 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [webhook approver]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":false,\\\"restartCount\\\":6,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:15Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.654559 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:15Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.667366 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.667409 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.667425 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.667445 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.667461 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:15Z","lastTransitionTime":"2026-03-12T00:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.667784 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfp2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"662dcbe7-7cf8-42d4-9a5e-345c32a49972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [egress-router-binary-copy cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfp2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:15Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.682001 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e483c2b-08f0-4e92-8e4a-b7281f30af3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [machine-config-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd9jv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd9jv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4xwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:15Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.698451 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5a9ae9441886817563a5fc793bff7c9b8d77d406ceb1d1fb1eebd4307a1c64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:15Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.711277 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwpqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b7580c2-5c23-4c67-807a-ea97a3df9398\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8lxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwpqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:15Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.725596 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-njc9c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0d6440c-4524-42ea-b18a-0ddc4662f672\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [node-ca]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7wn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-njc9c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:15Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.736781 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5a9ae9441886817563a5fc793bff7c9b8d77d406ceb1d1fb1eebd4307a1c64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:15Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.749234 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwpqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b7580c2-5c23-4c67-807a-ea97a3df9398\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d07c603619f858177d4b007e785208853abcfe2dc6c1af435309030fa409010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8lxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwpqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:15Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.763637 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-njc9c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0d6440c-4524-42ea-b18a-0ddc4662f672\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f4d0d9599ad62aaba35ba1414a806f3a163e29a3c34fd893ef5b14eae003591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7wn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-njc9c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:15Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.769110 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.769190 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.769206 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.769269 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.769282 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:15Z","lastTransitionTime":"2026-03-12T00:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.774196 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9s5fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f9c7f1-6593-4d92-854c-e6c9964a0435\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b8833cbaf3809463e18bfebc9d5bada5fb26823831cc742a3a27177662f60e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xf4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9s5fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:15Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.785841 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xdd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b5233d6-d2b8-42dc-b881-18e439ecc442\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0ea55d993c72b6d50a4829ecfff4adae7d0f625a90e40545c9e63201cdda29c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjpwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb23ff4bc484524e306561c0c574e2b2eee39fc80efcd9f44f1924fe42df1a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjpwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4xdd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:15Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.797874 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:15Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.810679 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:15Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.826210 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:15Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.836470 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mqmtd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be985ff1-de8a-431f-a230-67894ebe52d5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmsz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmsz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mqmtd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:15Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.860010 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15117613-9fad-48c7-98c4-a2d84502ded9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7mp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:15Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.871250 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5139686878eb07d3abced92d0ccfe05be8907580ed75233b72fc2ad3415a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15cbe3fa75b69a7176803295e16a314a8bc73f5797c874124d54e76d47fa1064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:15Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.871628 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.871669 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.871686 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.871707 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.871721 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:15Z","lastTransitionTime":"2026-03-12T00:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.885678 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:15Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.900916 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfp2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"662dcbe7-7cf8-42d4-9a5e-345c32a49972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [cni-plugins bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a57d476b5b00110168878bd1c19c32ef122525c0259ac0a70de31c8220996d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4a57d476b5b00110168878bd1c19c32ef122525c0259ac0a70de31c8220996d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfp2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:15Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.921831 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e483c2b-08f0-4e92-8e4a-b7281f30af3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d6839c99ab958ad1f40636c70e29bc285292fdbee1e910cdf7f8ab43ca7e8db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd9jv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82de8f2b633f1b6c295149d077bfd2b308f5acaa429acf411ed16f456536d0ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd9jv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4xwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:15Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.958851 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.958966 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:08:15 crc kubenswrapper[4948]: E0312 00:08:15.959033 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 00:08:17.959009126 +0000 UTC m=+97.414612864 (durationBeforeRetry 2s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:08:15 crc kubenswrapper[4948]: E0312 00:08:15.959058 4948 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.959092 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:08:15 crc kubenswrapper[4948]: E0312 00:08:15.959111 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-12 00:08:17.95909574 +0000 UTC m=+97.414699478 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.959143 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.959168 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:08:15 crc kubenswrapper[4948]: E0312 00:08:15.959298 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 12 00:08:15 crc kubenswrapper[4948]: E0312 00:08:15.959334 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 12 00:08:15 crc kubenswrapper[4948]: E0312 00:08:15.959345 4948 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 00:08:15 crc kubenswrapper[4948]: E0312 00:08:15.959384 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-12 00:08:17.959375834 +0000 UTC m=+97.414979572 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 00:08:15 crc kubenswrapper[4948]: E0312 00:08:15.959489 4948 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 12 00:08:15 crc kubenswrapper[4948]: E0312 00:08:15.959597 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 12 00:08:15 crc kubenswrapper[4948]: E0312 00:08:15.959669 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 12 00:08:15 crc kubenswrapper[4948]: E0312 00:08:15.959725 4948 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 00:08:15 crc kubenswrapper[4948]: E0312 00:08:15.959617 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-12 00:08:17.959590044 +0000 UTC m=+97.415193822 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 12 00:08:15 crc kubenswrapper[4948]: E0312 00:08:15.959853 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-12 00:08:17.959836955 +0000 UTC m=+97.415440723 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.974155 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.974201 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.974213 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.974229 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:15 crc kubenswrapper[4948]: I0312 00:08:15.974241 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:15Z","lastTransitionTime":"2026-03-12T00:08:15Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.060598 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/be985ff1-de8a-431f-a230-67894ebe52d5-metrics-certs\") pod \"network-metrics-daemon-mqmtd\" (UID: \"be985ff1-de8a-431f-a230-67894ebe52d5\") " pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:08:16 crc kubenswrapper[4948]: E0312 00:08:16.060743 4948 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 12 00:08:16 crc kubenswrapper[4948]: E0312 00:08:16.060796 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/be985ff1-de8a-431f-a230-67894ebe52d5-metrics-certs podName:be985ff1-de8a-431f-a230-67894ebe52d5 nodeName:}" failed. No retries permitted until 2026-03-12 00:08:18.060780376 +0000 UTC m=+97.516384114 (durationBeforeRetry 2s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/be985ff1-de8a-431f-a230-67894ebe52d5-metrics-certs") pod "network-metrics-daemon-mqmtd" (UID: "be985ff1-de8a-431f-a230-67894ebe52d5") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.076456 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.076497 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.076508 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.076523 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.076534 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:16Z","lastTransitionTime":"2026-03-12T00:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.179045 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.179594 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.179676 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.179758 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.179824 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:16Z","lastTransitionTime":"2026-03-12T00:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.281108 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.281334 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.281440 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.281527 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.281602 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:16Z","lastTransitionTime":"2026-03-12T00:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.312513 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.312517 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.312636 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.312857 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:08:16 crc kubenswrapper[4948]: E0312 00:08:16.312968 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 00:08:16 crc kubenswrapper[4948]: E0312 00:08:16.313078 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 00:08:16 crc kubenswrapper[4948]: E0312 00:08:16.313185 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mqmtd" podUID="be985ff1-de8a-431f-a230-67894ebe52d5" Mar 12 00:08:16 crc kubenswrapper[4948]: E0312 00:08:16.313270 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.384253 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.384295 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.384324 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.384341 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.384356 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:16Z","lastTransitionTime":"2026-03-12T00:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.490586 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.490622 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.490631 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.490647 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.490655 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:16Z","lastTransitionTime":"2026-03-12T00:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.547968 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" event={"ID":"15117613-9fad-48c7-98c4-a2d84502ded9","Type":"ContainerStarted","Data":"84af4cdd213d90636a889a8a8a3aaca3dee1e2677368efa0ffc768940c2cc994"} Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.548010 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" event={"ID":"15117613-9fad-48c7-98c4-a2d84502ded9","Type":"ContainerStarted","Data":"d4036d7131db00afd8eab8e093626b688297dc776b5af80c3c38b51da7124085"} Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.548021 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" event={"ID":"15117613-9fad-48c7-98c4-a2d84502ded9","Type":"ContainerStarted","Data":"ec9d5e77753fda70350353daa0967257dd4929dba4440a95495ab4c2a0659aef"} Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.548030 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" event={"ID":"15117613-9fad-48c7-98c4-a2d84502ded9","Type":"ContainerStarted","Data":"8cad943edb9a8d445b33a6e70d85b475c1ccba460a6056bccae94fe60fd6b73f"} Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.549351 4948 generic.go:334] "Generic (PLEG): container finished" podID="662dcbe7-7cf8-42d4-9a5e-345c32a49972" containerID="0bb429b94a1286d8aac77cb77112c7163f7aa50a1c9cab997f8d088c100887c2" exitCode=0 Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.549398 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-mfp2k" event={"ID":"662dcbe7-7cf8-42d4-9a5e-345c32a49972","Type":"ContainerDied","Data":"0bb429b94a1286d8aac77cb77112c7163f7aa50a1c9cab997f8d088c100887c2"} Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.565114 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5139686878eb07d3abced92d0ccfe05be8907580ed75233b72fc2ad3415a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15cbe3fa75b69a7176803295e16a314a8bc73f5797c874124d54e76d47fa1064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:16Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.588256 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:16Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.592686 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.592935 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.592955 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.592979 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.592996 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:16Z","lastTransitionTime":"2026-03-12T00:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.603124 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfp2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"662dcbe7-7cf8-42d4-9a5e-345c32a49972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [bond-cni-plugin routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a57d476b5b00110168878bd1c19c32ef122525c0259ac0a70de31c8220996d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4a57d476b5b00110168878bd1c19c32ef122525c0259ac0a70de31c8220996d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bb429b94a1286d8aac77cb77112c7163f7aa50a1c9cab997f8d088c100887c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0bb429b94a1286d8aac77cb77112c7163f7aa50a1c9cab997f8d088c100887c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfp2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:16Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.616040 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e483c2b-08f0-4e92-8e4a-b7281f30af3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d6839c99ab958ad1f40636c70e29bc285292fdbee1e910cdf7f8ab43ca7e8db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd9jv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82de8f2b633f1b6c295149d077bfd2b308f5acaa429acf411ed16f456536d0ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd9jv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4xwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:16Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.636516 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5a9ae9441886817563a5fc793bff7c9b8d77d406ceb1d1fb1eebd4307a1c64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:16Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.655717 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwpqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b7580c2-5c23-4c67-807a-ea97a3df9398\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d07c603619f858177d4b007e785208853abcfe2dc6c1af435309030fa409010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8lxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwpqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:16Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.668192 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-njc9c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0d6440c-4524-42ea-b18a-0ddc4662f672\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f4d0d9599ad62aaba35ba1414a806f3a163e29a3c34fd893ef5b14eae003591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7wn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-njc9c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:16Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.682559 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xdd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b5233d6-d2b8-42dc-b881-18e439ecc442\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0ea55d993c72b6d50a4829ecfff4adae7d0f625a90e40545c9e63201cdda29c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjpwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb23ff4bc484524e306561c0c574e2b2eee39fc80efcd9f44f1924fe42df1a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjpwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4xdd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:16Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.692929 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9s5fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f9c7f1-6593-4d92-854c-e6c9964a0435\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b8833cbaf3809463e18bfebc9d5bada5fb26823831cc742a3a27177662f60e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xf4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9s5fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:16Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.698231 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.698278 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.698293 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.698343 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.698357 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:16Z","lastTransitionTime":"2026-03-12T00:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.709228 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:16Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.731019 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mqmtd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be985ff1-de8a-431f-a230-67894ebe52d5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmsz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmsz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mqmtd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:16Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.754927 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15117613-9fad-48c7-98c4-a2d84502ded9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7mp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:16Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.772985 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:16Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.786535 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:16Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.801381 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.801436 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.801452 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.801477 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.801494 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:16Z","lastTransitionTime":"2026-03-12T00:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.904189 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.904247 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.904267 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.904323 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:16 crc kubenswrapper[4948]: I0312 00:08:16.904341 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:16Z","lastTransitionTime":"2026-03-12T00:08:16Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.006653 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.006684 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.006692 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.006706 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.006714 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:17Z","lastTransitionTime":"2026-03-12T00:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.108910 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.108946 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.108957 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.108970 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.108979 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:17Z","lastTransitionTime":"2026-03-12T00:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.210874 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.210921 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.210932 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.210949 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.210961 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:17Z","lastTransitionTime":"2026-03-12T00:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.313684 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.313725 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.313737 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.313755 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.313767 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:17Z","lastTransitionTime":"2026-03-12T00:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.416279 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.416363 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.416381 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.416409 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.416431 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:17Z","lastTransitionTime":"2026-03-12T00:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.520358 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.520430 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.520453 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.520485 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.520521 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:17Z","lastTransitionTime":"2026-03-12T00:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.558201 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" event={"ID":"15117613-9fad-48c7-98c4-a2d84502ded9","Type":"ContainerStarted","Data":"2e481ace9e84b6aa66cbdda3ee8425c4120a665915d2c70ae4faf60bba6205e2"} Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.558254 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" event={"ID":"15117613-9fad-48c7-98c4-a2d84502ded9","Type":"ContainerStarted","Data":"c9eadefd0babc119f776d48c9d2c9e4bced58db685af9a8e97f25e6a9a2e3c89"} Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.562919 4948 generic.go:334] "Generic (PLEG): container finished" podID="662dcbe7-7cf8-42d4-9a5e-345c32a49972" containerID="a77485fd9e246f190724b86663e0dbe81146e1fcc51f468dd05f23c1db878fd2" exitCode=0 Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.563058 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-mfp2k" event={"ID":"662dcbe7-7cf8-42d4-9a5e-345c32a49972","Type":"ContainerDied","Data":"a77485fd9e246f190724b86663e0dbe81146e1fcc51f468dd05f23c1db878fd2"} Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.566802 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" event={"ID":"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49","Type":"ContainerStarted","Data":"8c8d58ee45d75284e030700fa526b2c39197a26f6a391a7417f25a026f356b5b"} Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.587673 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5a9ae9441886817563a5fc793bff7c9b8d77d406ceb1d1fb1eebd4307a1c64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:17Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.610061 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwpqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b7580c2-5c23-4c67-807a-ea97a3df9398\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d07c603619f858177d4b007e785208853abcfe2dc6c1af435309030fa409010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8lxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwpqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:17Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.623434 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.623506 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.623524 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.623555 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.623573 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:17Z","lastTransitionTime":"2026-03-12T00:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.631819 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-njc9c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0d6440c-4524-42ea-b18a-0ddc4662f672\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f4d0d9599ad62aaba35ba1414a806f3a163e29a3c34fd893ef5b14eae003591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7wn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-njc9c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:17Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.647385 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xdd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b5233d6-d2b8-42dc-b881-18e439ecc442\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0ea55d993c72b6d50a4829ecfff4adae7d0f625a90e40545c9e63201cdda29c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjpwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb23ff4bc484524e306561c0c574e2b2eee39fc80efcd9f44f1924fe42df1a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjpwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4xdd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:17Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.662676 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9s5fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f9c7f1-6593-4d92-854c-e6c9964a0435\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b8833cbaf3809463e18bfebc9d5bada5fb26823831cc742a3a27177662f60e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xf4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9s5fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:17Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.679225 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [iptables-alerter]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:17Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.695460 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mqmtd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be985ff1-de8a-431f-a230-67894ebe52d5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmsz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmsz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mqmtd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:17Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.718324 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15117613-9fad-48c7-98c4-a2d84502ded9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7mp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:17Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.730529 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.730590 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.730608 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.730635 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.730652 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:17Z","lastTransitionTime":"2026-03-12T00:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.737917 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:17Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.758994 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:17Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.774393 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5139686878eb07d3abced92d0ccfe05be8907580ed75233b72fc2ad3415a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15cbe3fa75b69a7176803295e16a314a8bc73f5797c874124d54e76d47fa1064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:17Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.791438 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:17Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.813529 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfp2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"662dcbe7-7cf8-42d4-9a5e-345c32a49972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a57d476b5b00110168878bd1c19c32ef122525c0259ac0a70de31c8220996d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4a57d476b5b00110168878bd1c19c32ef122525c0259ac0a70de31c8220996d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bb429b94a1286d8aac77cb77112c7163f7aa50a1c9cab997f8d088c100887c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0bb429b94a1286d8aac77cb77112c7163f7aa50a1c9cab997f8d088c100887c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a77485fd9e246f190724b86663e0dbe81146e1fcc51f468dd05f23c1db878fd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a77485fd9e246f190724b86663e0dbe81146e1fcc51f468dd05f23c1db878fd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfp2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:17Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.832046 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e483c2b-08f0-4e92-8e4a-b7281f30af3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d6839c99ab958ad1f40636c70e29bc285292fdbee1e910cdf7f8ab43ca7e8db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd9jv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82de8f2b633f1b6c295149d077bfd2b308f5acaa429acf411ed16f456536d0ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd9jv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4xwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:17Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.835374 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.835448 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.835474 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.835512 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.835538 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:17Z","lastTransitionTime":"2026-03-12T00:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.845996 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e483c2b-08f0-4e92-8e4a-b7281f30af3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d6839c99ab958ad1f40636c70e29bc285292fdbee1e910cdf7f8ab43ca7e8db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd9jv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82de8f2b633f1b6c295149d077bfd2b308f5acaa429acf411ed16f456536d0ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd9jv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4xwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:17Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.863432 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5139686878eb07d3abced92d0ccfe05be8907580ed75233b72fc2ad3415a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15cbe3fa75b69a7176803295e16a314a8bc73f5797c874124d54e76d47fa1064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:17Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.882269 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:17Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.898595 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfp2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"662dcbe7-7cf8-42d4-9a5e-345c32a49972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [routeoverride-cni whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a57d476b5b00110168878bd1c19c32ef122525c0259ac0a70de31c8220996d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4a57d476b5b00110168878bd1c19c32ef122525c0259ac0a70de31c8220996d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bb429b94a1286d8aac77cb77112c7163f7aa50a1c9cab997f8d088c100887c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0bb429b94a1286d8aac77cb77112c7163f7aa50a1c9cab997f8d088c100887c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a77485fd9e246f190724b86663e0dbe81146e1fcc51f468dd05f23c1db878fd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a77485fd9e246f190724b86663e0dbe81146e1fcc51f468dd05f23c1db878fd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfp2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:17Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.919363 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5a9ae9441886817563a5fc793bff7c9b8d77d406ceb1d1fb1eebd4307a1c64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:17Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.933495 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwpqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b7580c2-5c23-4c67-807a-ea97a3df9398\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d07c603619f858177d4b007e785208853abcfe2dc6c1af435309030fa409010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8lxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwpqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:17Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.938250 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.938293 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.938328 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.938344 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.938354 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:17Z","lastTransitionTime":"2026-03-12T00:08:17Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.945502 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-njc9c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0d6440c-4524-42ea-b18a-0ddc4662f672\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f4d0d9599ad62aaba35ba1414a806f3a163e29a3c34fd893ef5b14eae003591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7wn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-njc9c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:17Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.956767 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9s5fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f9c7f1-6593-4d92-854c-e6c9964a0435\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b8833cbaf3809463e18bfebc9d5bada5fb26823831cc742a3a27177662f60e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xf4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9s5fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:17Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.968788 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xdd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b5233d6-d2b8-42dc-b881-18e439ecc442\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0ea55d993c72b6d50a4829ecfff4adae7d0f625a90e40545c9e63201cdda29c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjpwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb23ff4bc484524e306561c0c574e2b2eee39fc80efcd9f44f1924fe42df1a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjpwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4xdd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:17Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.981233 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:17Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.983455 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.983547 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.983573 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.983607 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.983627 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:08:17 crc kubenswrapper[4948]: E0312 00:08:17.983727 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 12 00:08:17 crc kubenswrapper[4948]: E0312 00:08:17.983741 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 12 00:08:17 crc kubenswrapper[4948]: E0312 00:08:17.983750 4948 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 00:08:17 crc kubenswrapper[4948]: E0312 00:08:17.983786 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-12 00:08:21.98377416 +0000 UTC m=+101.439377898 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 00:08:17 crc kubenswrapper[4948]: E0312 00:08:17.984027 4948 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 12 00:08:17 crc kubenswrapper[4948]: E0312 00:08:17.984053 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-12 00:08:21.984045983 +0000 UTC m=+101.439649721 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 12 00:08:17 crc kubenswrapper[4948]: E0312 00:08:17.984079 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 12 00:08:17 crc kubenswrapper[4948]: E0312 00:08:17.984099 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 00:08:21.984092445 +0000 UTC m=+101.439696183 (durationBeforeRetry 4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:08:17 crc kubenswrapper[4948]: E0312 00:08:17.984115 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 12 00:08:17 crc kubenswrapper[4948]: E0312 00:08:17.984125 4948 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 12 00:08:17 crc kubenswrapper[4948]: E0312 00:08:17.984136 4948 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 00:08:17 crc kubenswrapper[4948]: E0312 00:08:17.984145 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-12 00:08:21.984138967 +0000 UTC m=+101.439742705 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 12 00:08:17 crc kubenswrapper[4948]: E0312 00:08:17.984213 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-12 00:08:21.984190349 +0000 UTC m=+101.439794127 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 00:08:17 crc kubenswrapper[4948]: I0312 00:08:17.995217 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:17Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.011429 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8d58ee45d75284e030700fa526b2c39197a26f6a391a7417f25a026f356b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:18Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.024279 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mqmtd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be985ff1-de8a-431f-a230-67894ebe52d5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmsz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmsz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mqmtd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:18Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.041923 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.042000 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.042021 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.042103 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.042121 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:18Z","lastTransitionTime":"2026-03-12T00:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.052618 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15117613-9fad-48c7-98c4-a2d84502ded9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7mp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:18Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.085158 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/be985ff1-de8a-431f-a230-67894ebe52d5-metrics-certs\") pod \"network-metrics-daemon-mqmtd\" (UID: \"be985ff1-de8a-431f-a230-67894ebe52d5\") " pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:08:18 crc kubenswrapper[4948]: E0312 00:08:18.085391 4948 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 12 00:08:18 crc kubenswrapper[4948]: E0312 00:08:18.085483 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/be985ff1-de8a-431f-a230-67894ebe52d5-metrics-certs podName:be985ff1-de8a-431f-a230-67894ebe52d5 nodeName:}" failed. No retries permitted until 2026-03-12 00:08:22.085459555 +0000 UTC m=+101.541063323 (durationBeforeRetry 4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/be985ff1-de8a-431f-a230-67894ebe52d5-metrics-certs") pod "network-metrics-daemon-mqmtd" (UID: "be985ff1-de8a-431f-a230-67894ebe52d5") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.145268 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.145324 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.145338 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.145354 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.145365 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:18Z","lastTransitionTime":"2026-03-12T00:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.247969 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.248034 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.248053 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.248076 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.248094 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:18Z","lastTransitionTime":"2026-03-12T00:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.311758 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.311812 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.311821 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.311755 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:08:18 crc kubenswrapper[4948]: E0312 00:08:18.311956 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mqmtd" podUID="be985ff1-de8a-431f-a230-67894ebe52d5" Mar 12 00:08:18 crc kubenswrapper[4948]: E0312 00:08:18.312203 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 00:08:18 crc kubenswrapper[4948]: E0312 00:08:18.312277 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 00:08:18 crc kubenswrapper[4948]: E0312 00:08:18.312587 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.350891 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.350984 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.351006 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.351083 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.351112 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:18Z","lastTransitionTime":"2026-03-12T00:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.455210 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.455287 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.455356 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.455389 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.455411 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:18Z","lastTransitionTime":"2026-03-12T00:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.558937 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.558999 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.559019 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.559046 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.559069 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:18Z","lastTransitionTime":"2026-03-12T00:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.581841 4948 generic.go:334] "Generic (PLEG): container finished" podID="662dcbe7-7cf8-42d4-9a5e-345c32a49972" containerID="adaa20d9a4898057802a8c9759ab3cd485963180bfc4f6b926bf6d2c886bcf27" exitCode=0 Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.582441 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-mfp2k" event={"ID":"662dcbe7-7cf8-42d4-9a5e-345c32a49972","Type":"ContainerDied","Data":"adaa20d9a4898057802a8c9759ab3cd485963180bfc4f6b926bf6d2c886bcf27"} Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.609746 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e483c2b-08f0-4e92-8e4a-b7281f30af3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d6839c99ab958ad1f40636c70e29bc285292fdbee1e910cdf7f8ab43ca7e8db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd9jv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82de8f2b633f1b6c295149d077bfd2b308f5acaa429acf411ed16f456536d0ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd9jv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4xwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:18Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.631212 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5139686878eb07d3abced92d0ccfe05be8907580ed75233b72fc2ad3415a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15cbe3fa75b69a7176803295e16a314a8bc73f5797c874124d54e76d47fa1064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:18Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.650931 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:18Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.661091 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.661158 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.661176 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.661201 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.661218 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:18Z","lastTransitionTime":"2026-03-12T00:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.668272 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfp2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"662dcbe7-7cf8-42d4-9a5e-345c32a49972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni-bincopy whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a57d476b5b00110168878bd1c19c32ef122525c0259ac0a70de31c8220996d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4a57d476b5b00110168878bd1c19c32ef122525c0259ac0a70de31c8220996d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bb429b94a1286d8aac77cb77112c7163f7aa50a1c9cab997f8d088c100887c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0bb429b94a1286d8aac77cb77112c7163f7aa50a1c9cab997f8d088c100887c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a77485fd9e246f190724b86663e0dbe81146e1fcc51f468dd05f23c1db878fd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a77485fd9e246f190724b86663e0dbe81146e1fcc51f468dd05f23c1db878fd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adaa20d9a4898057802a8c9759ab3cd485963180bfc4f6b926bf6d2c886bcf27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adaa20d9a4898057802a8c9759ab3cd485963180bfc4f6b926bf6d2c886bcf27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfp2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:18Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.683852 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5a9ae9441886817563a5fc793bff7c9b8d77d406ceb1d1fb1eebd4307a1c64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:18Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.697954 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwpqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b7580c2-5c23-4c67-807a-ea97a3df9398\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d07c603619f858177d4b007e785208853abcfe2dc6c1af435309030fa409010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8lxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwpqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:18Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.709883 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-njc9c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0d6440c-4524-42ea-b18a-0ddc4662f672\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f4d0d9599ad62aaba35ba1414a806f3a163e29a3c34fd893ef5b14eae003591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7wn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-njc9c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:18Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.723387 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9s5fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f9c7f1-6593-4d92-854c-e6c9964a0435\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b8833cbaf3809463e18bfebc9d5bada5fb26823831cc742a3a27177662f60e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xf4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9s5fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:18Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.734996 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xdd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b5233d6-d2b8-42dc-b881-18e439ecc442\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0ea55d993c72b6d50a4829ecfff4adae7d0f625a90e40545c9e63201cdda29c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjpwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb23ff4bc484524e306561c0c574e2b2eee39fc80efcd9f44f1924fe42df1a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjpwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4xdd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:18Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.751223 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:18Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.763180 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:18Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.764519 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.764596 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.764610 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.764649 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.764684 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:18Z","lastTransitionTime":"2026-03-12T00:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.775630 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8d58ee45d75284e030700fa526b2c39197a26f6a391a7417f25a026f356b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:18Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.784202 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mqmtd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be985ff1-de8a-431f-a230-67894ebe52d5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmsz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmsz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mqmtd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:18Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.802485 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15117613-9fad-48c7-98c4-a2d84502ded9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7mp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:18Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.867278 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.867342 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.867359 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.867400 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.867416 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:18Z","lastTransitionTime":"2026-03-12T00:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.970065 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.970102 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.970110 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.970123 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:18 crc kubenswrapper[4948]: I0312 00:08:18.970133 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:18Z","lastTransitionTime":"2026-03-12T00:08:18Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:19 crc kubenswrapper[4948]: I0312 00:08:19.072545 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:19 crc kubenswrapper[4948]: I0312 00:08:19.072625 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:19 crc kubenswrapper[4948]: I0312 00:08:19.072644 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:19 crc kubenswrapper[4948]: I0312 00:08:19.072670 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:19 crc kubenswrapper[4948]: I0312 00:08:19.072690 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:19Z","lastTransitionTime":"2026-03-12T00:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:19 crc kubenswrapper[4948]: I0312 00:08:19.175709 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:19 crc kubenswrapper[4948]: I0312 00:08:19.175777 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:19 crc kubenswrapper[4948]: I0312 00:08:19.175796 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:19 crc kubenswrapper[4948]: I0312 00:08:19.175825 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:19 crc kubenswrapper[4948]: I0312 00:08:19.175850 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:19Z","lastTransitionTime":"2026-03-12T00:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:19 crc kubenswrapper[4948]: I0312 00:08:19.278602 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:19 crc kubenswrapper[4948]: I0312 00:08:19.278651 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:19 crc kubenswrapper[4948]: I0312 00:08:19.278661 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:19 crc kubenswrapper[4948]: I0312 00:08:19.278679 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:19 crc kubenswrapper[4948]: I0312 00:08:19.278690 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:19Z","lastTransitionTime":"2026-03-12T00:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:19 crc kubenswrapper[4948]: I0312 00:08:19.381143 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:19 crc kubenswrapper[4948]: I0312 00:08:19.381223 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:19 crc kubenswrapper[4948]: I0312 00:08:19.381240 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:19 crc kubenswrapper[4948]: I0312 00:08:19.381259 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:19 crc kubenswrapper[4948]: I0312 00:08:19.381273 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:19Z","lastTransitionTime":"2026-03-12T00:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:19 crc kubenswrapper[4948]: I0312 00:08:19.483494 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:19 crc kubenswrapper[4948]: I0312 00:08:19.483533 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:19 crc kubenswrapper[4948]: I0312 00:08:19.483545 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:19 crc kubenswrapper[4948]: I0312 00:08:19.483561 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:19 crc kubenswrapper[4948]: I0312 00:08:19.483574 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:19Z","lastTransitionTime":"2026-03-12T00:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:19 crc kubenswrapper[4948]: I0312 00:08:19.585436 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:19 crc kubenswrapper[4948]: I0312 00:08:19.585465 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:19 crc kubenswrapper[4948]: I0312 00:08:19.585473 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:19 crc kubenswrapper[4948]: I0312 00:08:19.585485 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:19 crc kubenswrapper[4948]: I0312 00:08:19.585494 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:19Z","lastTransitionTime":"2026-03-12T00:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:19 crc kubenswrapper[4948]: I0312 00:08:19.588166 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" event={"ID":"15117613-9fad-48c7-98c4-a2d84502ded9","Type":"ContainerStarted","Data":"46fa6140f4f4bd32e298fa36f6d7af6890029f9bc3080d2b4d94be8307346de8"} Mar 12 00:08:19 crc kubenswrapper[4948]: I0312 00:08:19.590065 4948 generic.go:334] "Generic (PLEG): container finished" podID="662dcbe7-7cf8-42d4-9a5e-345c32a49972" containerID="998fbb934f01c9ecd315f78e776199d4fd538106f8373f09b730d2df23ac9e3f" exitCode=0 Mar 12 00:08:19 crc kubenswrapper[4948]: I0312 00:08:19.590091 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-mfp2k" event={"ID":"662dcbe7-7cf8-42d4-9a5e-345c32a49972","Type":"ContainerDied","Data":"998fbb934f01c9ecd315f78e776199d4fd538106f8373f09b730d2df23ac9e3f"} Mar 12 00:08:19 crc kubenswrapper[4948]: I0312 00:08:19.608359 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e483c2b-08f0-4e92-8e4a-b7281f30af3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d6839c99ab958ad1f40636c70e29bc285292fdbee1e910cdf7f8ab43ca7e8db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd9jv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82de8f2b633f1b6c295149d077bfd2b308f5acaa429acf411ed16f456536d0ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd9jv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4xwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:19Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:19 crc kubenswrapper[4948]: I0312 00:08:19.629267 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5139686878eb07d3abced92d0ccfe05be8907580ed75233b72fc2ad3415a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15cbe3fa75b69a7176803295e16a314a8bc73f5797c874124d54e76d47fa1064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:19Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:19 crc kubenswrapper[4948]: I0312 00:08:19.646379 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:19Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:19 crc kubenswrapper[4948]: I0312 00:08:19.667379 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfp2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"662dcbe7-7cf8-42d4-9a5e-345c32a49972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with incomplete status: [whereabouts-cni]\\\",\\\"reason\\\":\\\"ContainersNotInitialized\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a57d476b5b00110168878bd1c19c32ef122525c0259ac0a70de31c8220996d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4a57d476b5b00110168878bd1c19c32ef122525c0259ac0a70de31c8220996d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bb429b94a1286d8aac77cb77112c7163f7aa50a1c9cab997f8d088c100887c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0bb429b94a1286d8aac77cb77112c7163f7aa50a1c9cab997f8d088c100887c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a77485fd9e246f190724b86663e0dbe81146e1fcc51f468dd05f23c1db878fd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a77485fd9e246f190724b86663e0dbe81146e1fcc51f468dd05f23c1db878fd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adaa20d9a4898057802a8c9759ab3cd485963180bfc4f6b926bf6d2c886bcf27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adaa20d9a4898057802a8c9759ab3cd485963180bfc4f6b926bf6d2c886bcf27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://998fbb934f01c9ecd315f78e776199d4fd538106f8373f09b730d2df23ac9e3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://998fbb934f01c9ecd315f78e776199d4fd538106f8373f09b730d2df23ac9e3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfp2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:19Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:19 crc kubenswrapper[4948]: I0312 00:08:19.685346 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5a9ae9441886817563a5fc793bff7c9b8d77d406ceb1d1fb1eebd4307a1c64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:19Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:19 crc kubenswrapper[4948]: I0312 00:08:19.689807 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:19 crc kubenswrapper[4948]: I0312 00:08:19.689835 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:19 crc kubenswrapper[4948]: I0312 00:08:19.689843 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:19 crc kubenswrapper[4948]: I0312 00:08:19.689857 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:19 crc kubenswrapper[4948]: I0312 00:08:19.689866 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:19Z","lastTransitionTime":"2026-03-12T00:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:19 crc kubenswrapper[4948]: I0312 00:08:19.705099 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwpqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b7580c2-5c23-4c67-807a-ea97a3df9398\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d07c603619f858177d4b007e785208853abcfe2dc6c1af435309030fa409010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8lxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwpqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:19Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:19 crc kubenswrapper[4948]: I0312 00:08:19.718802 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-njc9c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0d6440c-4524-42ea-b18a-0ddc4662f672\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f4d0d9599ad62aaba35ba1414a806f3a163e29a3c34fd893ef5b14eae003591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7wn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-njc9c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:19Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:19 crc kubenswrapper[4948]: I0312 00:08:19.733054 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9s5fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f9c7f1-6593-4d92-854c-e6c9964a0435\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b8833cbaf3809463e18bfebc9d5bada5fb26823831cc742a3a27177662f60e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xf4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9s5fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:19Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:19 crc kubenswrapper[4948]: I0312 00:08:19.749038 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xdd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b5233d6-d2b8-42dc-b881-18e439ecc442\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0ea55d993c72b6d50a4829ecfff4adae7d0f625a90e40545c9e63201cdda29c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjpwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb23ff4bc484524e306561c0c574e2b2eee39fc80efcd9f44f1924fe42df1a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjpwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4xdd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:19Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:19 crc kubenswrapper[4948]: I0312 00:08:19.766591 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:19Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:19 crc kubenswrapper[4948]: I0312 00:08:19.784623 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:19Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:19 crc kubenswrapper[4948]: I0312 00:08:19.795850 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:19 crc kubenswrapper[4948]: I0312 00:08:19.795876 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:19 crc kubenswrapper[4948]: I0312 00:08:19.795884 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:19 crc kubenswrapper[4948]: I0312 00:08:19.795899 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:19 crc kubenswrapper[4948]: I0312 00:08:19.795907 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:19Z","lastTransitionTime":"2026-03-12T00:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:19 crc kubenswrapper[4948]: I0312 00:08:19.799004 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8d58ee45d75284e030700fa526b2c39197a26f6a391a7417f25a026f356b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:19Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:19 crc kubenswrapper[4948]: I0312 00:08:19.811125 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mqmtd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be985ff1-de8a-431f-a230-67894ebe52d5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmsz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmsz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mqmtd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:19Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:19 crc kubenswrapper[4948]: I0312 00:08:19.842123 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15117613-9fad-48c7-98c4-a2d84502ded9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7mp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:19Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:19 crc kubenswrapper[4948]: I0312 00:08:19.899099 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:19 crc kubenswrapper[4948]: I0312 00:08:19.899150 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:19 crc kubenswrapper[4948]: I0312 00:08:19.899168 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:19 crc kubenswrapper[4948]: I0312 00:08:19.899191 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:19 crc kubenswrapper[4948]: I0312 00:08:19.899208 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:19Z","lastTransitionTime":"2026-03-12T00:08:19Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.003070 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.003122 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.003133 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.003153 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.003164 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:20Z","lastTransitionTime":"2026-03-12T00:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.106188 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.106233 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.106242 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.106256 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.106264 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:20Z","lastTransitionTime":"2026-03-12T00:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.209685 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.209735 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.209751 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.209772 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.209788 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:20Z","lastTransitionTime":"2026-03-12T00:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.311651 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:08:20 crc kubenswrapper[4948]: E0312 00:08:20.311847 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mqmtd" podUID="be985ff1-de8a-431f-a230-67894ebe52d5" Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.312490 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:08:20 crc kubenswrapper[4948]: E0312 00:08:20.312668 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.312752 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.312799 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:08:20 crc kubenswrapper[4948]: E0312 00:08:20.312872 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 00:08:20 crc kubenswrapper[4948]: E0312 00:08:20.313127 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.313968 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.314025 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.314047 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.314074 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.314095 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:20Z","lastTransitionTime":"2026-03-12T00:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.346797 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.347763 4948 scope.go:117] "RemoveContainer" containerID="73834c8db23f6536034dce09e38b56665211238280bf9a4e1b13e74d81245a77" Mar 12 00:08:20 crc kubenswrapper[4948]: E0312 00:08:20.348071 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.420761 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.420827 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.420844 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.420870 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.420887 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:20Z","lastTransitionTime":"2026-03-12T00:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.523036 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.523098 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.523115 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.523141 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.523158 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:20Z","lastTransitionTime":"2026-03-12T00:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.599466 4948 generic.go:334] "Generic (PLEG): container finished" podID="662dcbe7-7cf8-42d4-9a5e-345c32a49972" containerID="46ebe18804ff07eb0e922a6cc1c39ff93875302d231222642b211e3527fe6373" exitCode=0 Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.599661 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-mfp2k" event={"ID":"662dcbe7-7cf8-42d4-9a5e-345c32a49972","Type":"ContainerDied","Data":"46ebe18804ff07eb0e922a6cc1c39ff93875302d231222642b211e3527fe6373"} Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.600561 4948 scope.go:117] "RemoveContainer" containerID="73834c8db23f6536034dce09e38b56665211238280bf9a4e1b13e74d81245a77" Mar 12 00:08:20 crc kubenswrapper[4948]: E0312 00:08:20.600865 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-apiserver-check-endpoints\" with CrashLoopBackOff: \"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\"" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.622808 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9s5fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f9c7f1-6593-4d92-854c-e6c9964a0435\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b8833cbaf3809463e18bfebc9d5bada5fb26823831cc742a3a27177662f60e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xf4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9s5fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:20Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.626687 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.626737 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.626753 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.626780 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.626796 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:20Z","lastTransitionTime":"2026-03-12T00:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.651330 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xdd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b5233d6-d2b8-42dc-b881-18e439ecc442\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0ea55d993c72b6d50a4829ecfff4adae7d0f625a90e40545c9e63201cdda29c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjpwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb23ff4bc484524e306561c0c574e2b2eee39fc80efcd9f44f1924fe42df1a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjpwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4xdd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:20Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.672001 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:20Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.693982 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:20Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.708010 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8d58ee45d75284e030700fa526b2c39197a26f6a391a7417f25a026f356b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:20Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.719997 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mqmtd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be985ff1-de8a-431f-a230-67894ebe52d5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmsz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmsz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mqmtd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:20Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.729914 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.729976 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.729995 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.730019 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.730038 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:20Z","lastTransitionTime":"2026-03-12T00:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.749075 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15117613-9fad-48c7-98c4-a2d84502ded9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7mp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:20Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.766356 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5139686878eb07d3abced92d0ccfe05be8907580ed75233b72fc2ad3415a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15cbe3fa75b69a7176803295e16a314a8bc73f5797c874124d54e76d47fa1064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:20Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.779113 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:20Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.797578 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfp2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"662dcbe7-7cf8-42d4-9a5e-345c32a49972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a57d476b5b00110168878bd1c19c32ef122525c0259ac0a70de31c8220996d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4a57d476b5b00110168878bd1c19c32ef122525c0259ac0a70de31c8220996d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bb429b94a1286d8aac77cb77112c7163f7aa50a1c9cab997f8d088c100887c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0bb429b94a1286d8aac77cb77112c7163f7aa50a1c9cab997f8d088c100887c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a77485fd9e246f190724b86663e0dbe81146e1fcc51f468dd05f23c1db878fd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a77485fd9e246f190724b86663e0dbe81146e1fcc51f468dd05f23c1db878fd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adaa20d9a4898057802a8c9759ab3cd485963180bfc4f6b926bf6d2c886bcf27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adaa20d9a4898057802a8c9759ab3cd485963180bfc4f6b926bf6d2c886bcf27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://998fbb934f01c9ecd315f78e776199d4fd538106f8373f09b730d2df23ac9e3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://998fbb934f01c9ecd315f78e776199d4fd538106f8373f09b730d2df23ac9e3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46ebe18804ff07eb0e922a6cc1c39ff93875302d231222642b211e3527fe6373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46ebe18804ff07eb0e922a6cc1c39ff93875302d231222642b211e3527fe6373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfp2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:20Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.812678 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e483c2b-08f0-4e92-8e4a-b7281f30af3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d6839c99ab958ad1f40636c70e29bc285292fdbee1e910cdf7f8ab43ca7e8db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd9jv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82de8f2b633f1b6c295149d077bfd2b308f5acaa429acf411ed16f456536d0ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd9jv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4xwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:20Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.833749 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2efff53-8e12-41a1-bcb1-9a1d08c42a4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1cb328c8431b39dac735d49726a2eae40a367849b2c146bb2e49d02958f0e5a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ad13662849131222d5c8ba1041b81fa75bcb1fcbfa2e71c0c751f04b532a5b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8cd4faf47b25395ffd42d4bf5416ea29dcaa3ec380d6d6d6a9f13321d35195a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73834c8db23f6536034dce09e38b56665211238280bf9a4e1b13e74d81245a77\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73834c8db23f6536034dce09e38b56665211238280bf9a4e1b13e74d81245a77\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T00:07:50Z\\\",\\\"message\\\":\\\"le observer\\\\nW0312 00:07:49.927227 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0312 00:07:49.927347 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 00:07:49.928000 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2511358960/tls.crt::/tmp/serving-cert-2511358960/tls.key\\\\\\\"\\\\nI0312 00:07:50.321455 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0312 00:07:50.324479 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0312 00:07:50.324511 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0312 00:07:50.324546 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0312 00:07:50.324556 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0312 00:07:50.331591 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0312 00:07:50.331609 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0312 00:07:50.331634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 00:07:50.331642 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 00:07:50.331651 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0312 00:07:50.331658 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0312 00:07:50.331664 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0312 00:07:50.331670 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0312 00:07:50.332917 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:07:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d56e32f9cb96ad103bab80d6a730b70d3387264a25a7a61ef1f0726e7290d6c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e42f2e94463ac8a3d7980a25122062da9ee87fcdc1f10e260ff9974b2e42359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e42f2e94463ac8a3d7980a25122062da9ee87fcdc1f10e260ff9974b2e42359\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:06:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:20Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.834706 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.834730 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.834740 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.834755 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.834766 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:20Z","lastTransitionTime":"2026-03-12T00:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.851829 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5a9ae9441886817563a5fc793bff7c9b8d77d406ceb1d1fb1eebd4307a1c64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:20Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.868017 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwpqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b7580c2-5c23-4c67-807a-ea97a3df9398\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d07c603619f858177d4b007e785208853abcfe2dc6c1af435309030fa409010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8lxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwpqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:20Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.881880 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-njc9c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0d6440c-4524-42ea-b18a-0ddc4662f672\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f4d0d9599ad62aaba35ba1414a806f3a163e29a3c34fd893ef5b14eae003591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7wn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-njc9c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:20Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.937668 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.937745 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.937776 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.937802 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:20 crc kubenswrapper[4948]: I0312 00:08:20.937821 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:20Z","lastTransitionTime":"2026-03-12T00:08:20Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.040959 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.041003 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.041012 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.041025 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.041035 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:21Z","lastTransitionTime":"2026-03-12T00:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.143549 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.143607 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.143620 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.143639 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.143651 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:21Z","lastTransitionTime":"2026-03-12T00:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.246903 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.246960 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.246976 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.247000 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.247030 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:21Z","lastTransitionTime":"2026-03-12T00:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.329214 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9s5fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f9c7f1-6593-4d92-854c-e6c9964a0435\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b8833cbaf3809463e18bfebc9d5bada5fb26823831cc742a3a27177662f60e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xf4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9s5fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:21Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.346784 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xdd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b5233d6-d2b8-42dc-b881-18e439ecc442\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0ea55d993c72b6d50a4829ecfff4adae7d0f625a90e40545c9e63201cdda29c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjpwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb23ff4bc484524e306561c0c574e2b2eee39fc80efcd9f44f1924fe42df1a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjpwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4xdd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:21Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.349657 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.349921 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.350046 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.350431 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.350615 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:21Z","lastTransitionTime":"2026-03-12T00:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.367986 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:21Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.388397 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:21Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.410041 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8d58ee45d75284e030700fa526b2c39197a26f6a391a7417f25a026f356b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:21Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.427496 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mqmtd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be985ff1-de8a-431f-a230-67894ebe52d5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmsz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmsz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mqmtd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:21Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.453412 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.453477 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.453496 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.453526 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.453550 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:21Z","lastTransitionTime":"2026-03-12T00:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.468382 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15117613-9fad-48c7-98c4-a2d84502ded9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7mp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:21Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.484819 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5139686878eb07d3abced92d0ccfe05be8907580ed75233b72fc2ad3415a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15cbe3fa75b69a7176803295e16a314a8bc73f5797c874124d54e76d47fa1064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:21Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.500708 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:21Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.525197 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfp2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"662dcbe7-7cf8-42d4-9a5e-345c32a49972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus-additional-cni-plugins]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a57d476b5b00110168878bd1c19c32ef122525c0259ac0a70de31c8220996d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4a57d476b5b00110168878bd1c19c32ef122525c0259ac0a70de31c8220996d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bb429b94a1286d8aac77cb77112c7163f7aa50a1c9cab997f8d088c100887c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0bb429b94a1286d8aac77cb77112c7163f7aa50a1c9cab997f8d088c100887c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a77485fd9e246f190724b86663e0dbe81146e1fcc51f468dd05f23c1db878fd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a77485fd9e246f190724b86663e0dbe81146e1fcc51f468dd05f23c1db878fd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adaa20d9a4898057802a8c9759ab3cd485963180bfc4f6b926bf6d2c886bcf27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adaa20d9a4898057802a8c9759ab3cd485963180bfc4f6b926bf6d2c886bcf27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://998fbb934f01c9ecd315f78e776199d4fd538106f8373f09b730d2df23ac9e3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://998fbb934f01c9ecd315f78e776199d4fd538106f8373f09b730d2df23ac9e3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46ebe18804ff07eb0e922a6cc1c39ff93875302d231222642b211e3527fe6373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46ebe18804ff07eb0e922a6cc1c39ff93875302d231222642b211e3527fe6373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfp2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:21Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.539625 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e483c2b-08f0-4e92-8e4a-b7281f30af3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d6839c99ab958ad1f40636c70e29bc285292fdbee1e910cdf7f8ab43ca7e8db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd9jv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82de8f2b633f1b6c295149d077bfd2b308f5acaa429acf411ed16f456536d0ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd9jv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4xwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:21Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.553293 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2efff53-8e12-41a1-bcb1-9a1d08c42a4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1cb328c8431b39dac735d49726a2eae40a367849b2c146bb2e49d02958f0e5a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ad13662849131222d5c8ba1041b81fa75bcb1fcbfa2e71c0c751f04b532a5b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8cd4faf47b25395ffd42d4bf5416ea29dcaa3ec380d6d6d6a9f13321d35195a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73834c8db23f6536034dce09e38b56665211238280bf9a4e1b13e74d81245a77\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73834c8db23f6536034dce09e38b56665211238280bf9a4e1b13e74d81245a77\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T00:07:50Z\\\",\\\"message\\\":\\\"le observer\\\\nW0312 00:07:49.927227 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0312 00:07:49.927347 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 00:07:49.928000 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2511358960/tls.crt::/tmp/serving-cert-2511358960/tls.key\\\\\\\"\\\\nI0312 00:07:50.321455 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0312 00:07:50.324479 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0312 00:07:50.324511 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0312 00:07:50.324546 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0312 00:07:50.324556 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0312 00:07:50.331591 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0312 00:07:50.331609 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0312 00:07:50.331634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 00:07:50.331642 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 00:07:50.331651 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0312 00:07:50.331658 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0312 00:07:50.331664 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0312 00:07:50.331670 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0312 00:07:50.332917 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:07:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d56e32f9cb96ad103bab80d6a730b70d3387264a25a7a61ef1f0726e7290d6c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e42f2e94463ac8a3d7980a25122062da9ee87fcdc1f10e260ff9974b2e42359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e42f2e94463ac8a3d7980a25122062da9ee87fcdc1f10e260ff9974b2e42359\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:06:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:21Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.556035 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.556092 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.556110 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.556132 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.556147 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:21Z","lastTransitionTime":"2026-03-12T00:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.565065 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5a9ae9441886817563a5fc793bff7c9b8d77d406ceb1d1fb1eebd4307a1c64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:21Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.580606 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwpqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b7580c2-5c23-4c67-807a-ea97a3df9398\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d07c603619f858177d4b007e785208853abcfe2dc6c1af435309030fa409010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8lxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwpqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:21Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.591835 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-njc9c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0d6440c-4524-42ea-b18a-0ddc4662f672\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f4d0d9599ad62aaba35ba1414a806f3a163e29a3c34fd893ef5b14eae003591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7wn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-njc9c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:21Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.608393 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-additional-cni-plugins-mfp2k" event={"ID":"662dcbe7-7cf8-42d4-9a5e-345c32a49972","Type":"ContainerStarted","Data":"ff3f3c69dcb28c7b0fd85f9e0c33b3339a1f6159dbf10bcf0efbdce4f7728c86"} Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.614327 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" event={"ID":"15117613-9fad-48c7-98c4-a2d84502ded9","Type":"ContainerStarted","Data":"1610d74076d215d3451c6503afb264abd468647b65c5da3e2cf8983ed1aeaeb2"} Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.615165 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.615232 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.615256 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.625043 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9s5fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f9c7f1-6593-4d92-854c-e6c9964a0435\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b8833cbaf3809463e18bfebc9d5bada5fb26823831cc742a3a27177662f60e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xf4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9s5fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:21Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.644475 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xdd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b5233d6-d2b8-42dc-b881-18e439ecc442\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0ea55d993c72b6d50a4829ecfff4adae7d0f625a90e40545c9e63201cdda29c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjpwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb23ff4bc484524e306561c0c574e2b2eee39fc80efcd9f44f1924fe42df1a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjpwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4xdd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:21Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.645422 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.648968 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.658634 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.658672 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.658681 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.658694 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.658703 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:21Z","lastTransitionTime":"2026-03-12T00:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.664862 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:21Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.675461 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:21Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.685754 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8d58ee45d75284e030700fa526b2c39197a26f6a391a7417f25a026f356b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:21Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.695677 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.695748 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.695760 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.695777 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.695788 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:21Z","lastTransitionTime":"2026-03-12T00:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.697830 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mqmtd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be985ff1-de8a-431f-a230-67894ebe52d5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmsz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmsz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mqmtd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:21Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:21 crc kubenswrapper[4948]: E0312 00:08:21.711985 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17825588-29cf-4059-bab7-1b042b9bf1f3\\\",\\\"systemUUID\\\":\\\"b18e4bf2-d37e-4b80-8da3-cd3803754ba2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:21Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.714887 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.714922 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.714931 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.714945 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.714955 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:21Z","lastTransitionTime":"2026-03-12T00:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.716424 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15117613-9fad-48c7-98c4-a2d84502ded9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovn-controller ovn-acl-logging kube-rbac-proxy-node kube-rbac-proxy-ovn-metrics northd nbdb sbdb ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"PodInitializing\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7mp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:21Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.725694 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e483c2b-08f0-4e92-8e4a-b7281f30af3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d6839c99ab958ad1f40636c70e29bc285292fdbee1e910cdf7f8ab43ca7e8db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd9jv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82de8f2b633f1b6c295149d077bfd2b308f5acaa429acf411ed16f456536d0ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd9jv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4xwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:21Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:21 crc kubenswrapper[4948]: E0312 00:08:21.727912 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17825588-29cf-4059-bab7-1b042b9bf1f3\\\",\\\"systemUUID\\\":\\\"b18e4bf2-d37e-4b80-8da3-cd3803754ba2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:21Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.731735 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.731790 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.731808 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.731829 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.731844 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:21Z","lastTransitionTime":"2026-03-12T00:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.738413 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5139686878eb07d3abced92d0ccfe05be8907580ed75233b72fc2ad3415a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15cbe3fa75b69a7176803295e16a314a8bc73f5797c874124d54e76d47fa1064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:21Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:21 crc kubenswrapper[4948]: E0312 00:08:21.745172 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17825588-29cf-4059-bab7-1b042b9bf1f3\\\",\\\"systemUUID\\\":\\\"b18e4bf2-d37e-4b80-8da3-cd3803754ba2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:21Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.748918 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.748944 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.748951 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.748964 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.748974 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:21Z","lastTransitionTime":"2026-03-12T00:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.751815 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:21Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:21 crc kubenswrapper[4948]: E0312 00:08:21.758957 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17825588-29cf-4059-bab7-1b042b9bf1f3\\\",\\\"systemUUID\\\":\\\"b18e4bf2-d37e-4b80-8da3-cd3803754ba2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:21Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.761705 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.761736 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.761745 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.761758 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.761766 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:21Z","lastTransitionTime":"2026-03-12T00:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.770997 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfp2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"662dcbe7-7cf8-42d4-9a5e-345c32a49972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff3f3c69dcb28c7b0fd85f9e0c33b3339a1f6159dbf10bcf0efbdce4f7728c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a57d476b5b00110168878bd1c19c32ef122525c0259ac0a70de31c8220996d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4a57d476b5b00110168878bd1c19c32ef122525c0259ac0a70de31c8220996d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bb429b94a1286d8aac77cb77112c7163f7aa50a1c9cab997f8d088c100887c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0bb429b94a1286d8aac77cb77112c7163f7aa50a1c9cab997f8d088c100887c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a77485fd9e246f190724b86663e0dbe81146e1fcc51f468dd05f23c1db878fd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a77485fd9e246f190724b86663e0dbe81146e1fcc51f468dd05f23c1db878fd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adaa20d9a4898057802a8c9759ab3cd485963180bfc4f6b926bf6d2c886bcf27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adaa20d9a4898057802a8c9759ab3cd485963180bfc4f6b926bf6d2c886bcf27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://998fbb934f01c9ecd315f78e776199d4fd538106f8373f09b730d2df23ac9e3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://998fbb934f01c9ecd315f78e776199d4fd538106f8373f09b730d2df23ac9e3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46ebe18804ff07eb0e922a6cc1c39ff93875302d231222642b211e3527fe6373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46ebe18804ff07eb0e922a6cc1c39ff93875302d231222642b211e3527fe6373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfp2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:21Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:21 crc kubenswrapper[4948]: E0312 00:08:21.777732 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17825588-29cf-4059-bab7-1b042b9bf1f3\\\",\\\"systemUUID\\\":\\\"b18e4bf2-d37e-4b80-8da3-cd3803754ba2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:21Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:21 crc kubenswrapper[4948]: E0312 00:08:21.777961 4948 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.779748 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.779892 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.780047 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.780192 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.780342 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:21Z","lastTransitionTime":"2026-03-12T00:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.789361 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2efff53-8e12-41a1-bcb1-9a1d08c42a4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1cb328c8431b39dac735d49726a2eae40a367849b2c146bb2e49d02958f0e5a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ad13662849131222d5c8ba1041b81fa75bcb1fcbfa2e71c0c751f04b532a5b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8cd4faf47b25395ffd42d4bf5416ea29dcaa3ec380d6d6d6a9f13321d35195a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73834c8db23f6536034dce09e38b56665211238280bf9a4e1b13e74d81245a77\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73834c8db23f6536034dce09e38b56665211238280bf9a4e1b13e74d81245a77\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T00:07:50Z\\\",\\\"message\\\":\\\"le observer\\\\nW0312 00:07:49.927227 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0312 00:07:49.927347 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 00:07:49.928000 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2511358960/tls.crt::/tmp/serving-cert-2511358960/tls.key\\\\\\\"\\\\nI0312 00:07:50.321455 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0312 00:07:50.324479 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0312 00:07:50.324511 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0312 00:07:50.324546 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0312 00:07:50.324556 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0312 00:07:50.331591 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0312 00:07:50.331609 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0312 00:07:50.331634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 00:07:50.331642 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 00:07:50.331651 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0312 00:07:50.331658 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0312 00:07:50.331664 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0312 00:07:50.331670 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0312 00:07:50.332917 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:07:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d56e32f9cb96ad103bab80d6a730b70d3387264a25a7a61ef1f0726e7290d6c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e42f2e94463ac8a3d7980a25122062da9ee87fcdc1f10e260ff9974b2e42359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e42f2e94463ac8a3d7980a25122062da9ee87fcdc1f10e260ff9974b2e42359\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:06:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:21Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.800513 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5a9ae9441886817563a5fc793bff7c9b8d77d406ceb1d1fb1eebd4307a1c64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:21Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.812335 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwpqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b7580c2-5c23-4c67-807a-ea97a3df9398\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d07c603619f858177d4b007e785208853abcfe2dc6c1af435309030fa409010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8lxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwpqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:21Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.820095 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-njc9c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0d6440c-4524-42ea-b18a-0ddc4662f672\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f4d0d9599ad62aaba35ba1414a806f3a163e29a3c34fd893ef5b14eae003591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7wn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-njc9c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:21Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.829938 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-njc9c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0d6440c-4524-42ea-b18a-0ddc4662f672\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f4d0d9599ad62aaba35ba1414a806f3a163e29a3c34fd893ef5b14eae003591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7wn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-njc9c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:21Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.844569 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2efff53-8e12-41a1-bcb1-9a1d08c42a4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1cb328c8431b39dac735d49726a2eae40a367849b2c146bb2e49d02958f0e5a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ad13662849131222d5c8ba1041b81fa75bcb1fcbfa2e71c0c751f04b532a5b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8cd4faf47b25395ffd42d4bf5416ea29dcaa3ec380d6d6d6a9f13321d35195a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73834c8db23f6536034dce09e38b56665211238280bf9a4e1b13e74d81245a77\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73834c8db23f6536034dce09e38b56665211238280bf9a4e1b13e74d81245a77\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T00:07:50Z\\\",\\\"message\\\":\\\"le observer\\\\nW0312 00:07:49.927227 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0312 00:07:49.927347 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 00:07:49.928000 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2511358960/tls.crt::/tmp/serving-cert-2511358960/tls.key\\\\\\\"\\\\nI0312 00:07:50.321455 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0312 00:07:50.324479 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0312 00:07:50.324511 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0312 00:07:50.324546 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0312 00:07:50.324556 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0312 00:07:50.331591 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0312 00:07:50.331609 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0312 00:07:50.331634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 00:07:50.331642 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 00:07:50.331651 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0312 00:07:50.331658 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0312 00:07:50.331664 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0312 00:07:50.331670 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0312 00:07:50.332917 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:07:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d56e32f9cb96ad103bab80d6a730b70d3387264a25a7a61ef1f0726e7290d6c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e42f2e94463ac8a3d7980a25122062da9ee87fcdc1f10e260ff9974b2e42359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e42f2e94463ac8a3d7980a25122062da9ee87fcdc1f10e260ff9974b2e42359\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:06:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:21Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.856659 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5a9ae9441886817563a5fc793bff7c9b8d77d406ceb1d1fb1eebd4307a1c64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:21Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.867617 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwpqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b7580c2-5c23-4c67-807a-ea97a3df9398\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d07c603619f858177d4b007e785208853abcfe2dc6c1af435309030fa409010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8lxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwpqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:21Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.876665 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9s5fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f9c7f1-6593-4d92-854c-e6c9964a0435\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b8833cbaf3809463e18bfebc9d5bada5fb26823831cc742a3a27177662f60e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xf4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9s5fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:21Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.882518 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.882558 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.882568 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.882582 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.882593 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:21Z","lastTransitionTime":"2026-03-12T00:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.890613 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xdd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b5233d6-d2b8-42dc-b881-18e439ecc442\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0ea55d993c72b6d50a4829ecfff4adae7d0f625a90e40545c9e63201cdda29c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjpwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb23ff4bc484524e306561c0c574e2b2eee39fc80efcd9f44f1924fe42df1a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjpwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4xdd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:21Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.915445 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15117613-9fad-48c7-98c4-a2d84502ded9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4036d7131db00afd8eab8e093626b688297dc776b5af80c3c38b51da7124085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84af4cdd213d90636a889a8a8a3aaca3dee1e2677368efa0ffc768940c2cc994\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e481ace9e84b6aa66cbdda3ee8425c4120a665915d2c70ae4faf60bba6205e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9eadefd0babc119f776d48c9d2c9e4bced58db685af9a8e97f25e6a9a2e3c89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9d5e77753fda70350353daa0967257dd4929dba4440a95495ab4c2a0659aef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cad943edb9a8d445b33a6e70d85b475c1ccba460a6056bccae94fe60fd6b73f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1610d74076d215d3451c6503afb264abd468647b65c5da3e2cf8983ed1aeaeb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46fa6140f4f4bd32e298fa36f6d7af6890029f9bc3080d2b4d94be8307346de8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7mp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:21Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.937185 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:21Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.959178 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:21Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.977740 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8d58ee45d75284e030700fa526b2c39197a26f6a391a7417f25a026f356b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:21Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.985124 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.985171 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.985190 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.985212 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.985228 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:21Z","lastTransitionTime":"2026-03-12T00:08:21Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:21 crc kubenswrapper[4948]: I0312 00:08:21.989175 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mqmtd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be985ff1-de8a-431f-a230-67894ebe52d5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmsz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmsz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mqmtd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:21Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:22 crc kubenswrapper[4948]: I0312 00:08:22.005351 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfp2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"662dcbe7-7cf8-42d4-9a5e-345c32a49972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff3f3c69dcb28c7b0fd85f9e0c33b3339a1f6159dbf10bcf0efbdce4f7728c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a57d476b5b00110168878bd1c19c32ef122525c0259ac0a70de31c8220996d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4a57d476b5b00110168878bd1c19c32ef122525c0259ac0a70de31c8220996d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bb429b94a1286d8aac77cb77112c7163f7aa50a1c9cab997f8d088c100887c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0bb429b94a1286d8aac77cb77112c7163f7aa50a1c9cab997f8d088c100887c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a77485fd9e246f190724b86663e0dbe81146e1fcc51f468dd05f23c1db878fd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a77485fd9e246f190724b86663e0dbe81146e1fcc51f468dd05f23c1db878fd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adaa20d9a4898057802a8c9759ab3cd485963180bfc4f6b926bf6d2c886bcf27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adaa20d9a4898057802a8c9759ab3cd485963180bfc4f6b926bf6d2c886bcf27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://998fbb934f01c9ecd315f78e776199d4fd538106f8373f09b730d2df23ac9e3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://998fbb934f01c9ecd315f78e776199d4fd538106f8373f09b730d2df23ac9e3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46ebe18804ff07eb0e922a6cc1c39ff93875302d231222642b211e3527fe6373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46ebe18804ff07eb0e922a6cc1c39ff93875302d231222642b211e3527fe6373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfp2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:22Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:22 crc kubenswrapper[4948]: I0312 00:08:22.016102 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e483c2b-08f0-4e92-8e4a-b7281f30af3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d6839c99ab958ad1f40636c70e29bc285292fdbee1e910cdf7f8ab43ca7e8db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd9jv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82de8f2b633f1b6c295149d077bfd2b308f5acaa429acf411ed16f456536d0ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd9jv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4xwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:22Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:22 crc kubenswrapper[4948]: I0312 00:08:22.028695 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 00:08:22 crc kubenswrapper[4948]: E0312 00:08:22.028885 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 00:08:30.028865735 +0000 UTC m=+109.484469483 (durationBeforeRetry 8s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:08:22 crc kubenswrapper[4948]: I0312 00:08:22.028925 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:08:22 crc kubenswrapper[4948]: I0312 00:08:22.028957 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:08:22 crc kubenswrapper[4948]: I0312 00:08:22.028995 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:08:22 crc kubenswrapper[4948]: I0312 00:08:22.029020 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:08:22 crc kubenswrapper[4948]: E0312 00:08:22.029119 4948 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 12 00:08:22 crc kubenswrapper[4948]: E0312 00:08:22.029163 4948 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 12 00:08:22 crc kubenswrapper[4948]: E0312 00:08:22.029146 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 12 00:08:22 crc kubenswrapper[4948]: E0312 00:08:22.029227 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 12 00:08:22 crc kubenswrapper[4948]: E0312 00:08:22.029272 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 12 00:08:22 crc kubenswrapper[4948]: E0312 00:08:22.029233 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-12 00:08:30.029212281 +0000 UTC m=+109.484816029 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 12 00:08:22 crc kubenswrapper[4948]: E0312 00:08:22.029297 4948 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 00:08:22 crc kubenswrapper[4948]: E0312 00:08:22.029354 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 12 00:08:22 crc kubenswrapper[4948]: E0312 00:08:22.029381 4948 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 00:08:22 crc kubenswrapper[4948]: E0312 00:08:22.029418 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-12 00:08:30.029358188 +0000 UTC m=+109.484961936 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 12 00:08:22 crc kubenswrapper[4948]: E0312 00:08:22.029446 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-12 00:08:30.029435872 +0000 UTC m=+109.485039740 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 00:08:22 crc kubenswrapper[4948]: E0312 00:08:22.029464 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-12 00:08:30.029456042 +0000 UTC m=+109.485059911 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 00:08:22 crc kubenswrapper[4948]: I0312 00:08:22.036366 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5139686878eb07d3abced92d0ccfe05be8907580ed75233b72fc2ad3415a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15cbe3fa75b69a7176803295e16a314a8bc73f5797c874124d54e76d47fa1064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:22Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:22 crc kubenswrapper[4948]: I0312 00:08:22.048393 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:22Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:22 crc kubenswrapper[4948]: I0312 00:08:22.087241 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:22 crc kubenswrapper[4948]: I0312 00:08:22.087278 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:22 crc kubenswrapper[4948]: I0312 00:08:22.087289 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:22 crc kubenswrapper[4948]: I0312 00:08:22.087319 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:22 crc kubenswrapper[4948]: I0312 00:08:22.087331 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:22Z","lastTransitionTime":"2026-03-12T00:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:22 crc kubenswrapper[4948]: I0312 00:08:22.130330 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/be985ff1-de8a-431f-a230-67894ebe52d5-metrics-certs\") pod \"network-metrics-daemon-mqmtd\" (UID: \"be985ff1-de8a-431f-a230-67894ebe52d5\") " pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:08:22 crc kubenswrapper[4948]: E0312 00:08:22.130504 4948 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 12 00:08:22 crc kubenswrapper[4948]: E0312 00:08:22.130577 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/be985ff1-de8a-431f-a230-67894ebe52d5-metrics-certs podName:be985ff1-de8a-431f-a230-67894ebe52d5 nodeName:}" failed. No retries permitted until 2026-03-12 00:08:30.13055949 +0000 UTC m=+109.586163218 (durationBeforeRetry 8s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/be985ff1-de8a-431f-a230-67894ebe52d5-metrics-certs") pod "network-metrics-daemon-mqmtd" (UID: "be985ff1-de8a-431f-a230-67894ebe52d5") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 12 00:08:22 crc kubenswrapper[4948]: I0312 00:08:22.189988 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:22 crc kubenswrapper[4948]: I0312 00:08:22.190042 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:22 crc kubenswrapper[4948]: I0312 00:08:22.190059 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:22 crc kubenswrapper[4948]: I0312 00:08:22.190083 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:22 crc kubenswrapper[4948]: I0312 00:08:22.190101 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:22Z","lastTransitionTime":"2026-03-12T00:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:22 crc kubenswrapper[4948]: I0312 00:08:22.293283 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:22 crc kubenswrapper[4948]: I0312 00:08:22.293390 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:22 crc kubenswrapper[4948]: I0312 00:08:22.293408 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:22 crc kubenswrapper[4948]: I0312 00:08:22.293433 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:22 crc kubenswrapper[4948]: I0312 00:08:22.293452 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:22Z","lastTransitionTime":"2026-03-12T00:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:22 crc kubenswrapper[4948]: I0312 00:08:22.312257 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:08:22 crc kubenswrapper[4948]: I0312 00:08:22.312412 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:08:22 crc kubenswrapper[4948]: I0312 00:08:22.312434 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:08:22 crc kubenswrapper[4948]: E0312 00:08:22.312499 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 00:08:22 crc kubenswrapper[4948]: I0312 00:08:22.312533 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:08:22 crc kubenswrapper[4948]: E0312 00:08:22.312688 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mqmtd" podUID="be985ff1-de8a-431f-a230-67894ebe52d5" Mar 12 00:08:22 crc kubenswrapper[4948]: E0312 00:08:22.312801 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 00:08:22 crc kubenswrapper[4948]: E0312 00:08:22.312891 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 00:08:22 crc kubenswrapper[4948]: I0312 00:08:22.396614 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:22 crc kubenswrapper[4948]: I0312 00:08:22.396666 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:22 crc kubenswrapper[4948]: I0312 00:08:22.396683 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:22 crc kubenswrapper[4948]: I0312 00:08:22.396707 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:22 crc kubenswrapper[4948]: I0312 00:08:22.396724 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:22Z","lastTransitionTime":"2026-03-12T00:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:22 crc kubenswrapper[4948]: I0312 00:08:22.500091 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:22 crc kubenswrapper[4948]: I0312 00:08:22.500152 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:22 crc kubenswrapper[4948]: I0312 00:08:22.500170 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:22 crc kubenswrapper[4948]: I0312 00:08:22.500196 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:22 crc kubenswrapper[4948]: I0312 00:08:22.500217 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:22Z","lastTransitionTime":"2026-03-12T00:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:22 crc kubenswrapper[4948]: I0312 00:08:22.603854 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:22 crc kubenswrapper[4948]: I0312 00:08:22.603924 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:22 crc kubenswrapper[4948]: I0312 00:08:22.603942 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:22 crc kubenswrapper[4948]: I0312 00:08:22.603966 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:22 crc kubenswrapper[4948]: I0312 00:08:22.603984 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:22Z","lastTransitionTime":"2026-03-12T00:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:22 crc kubenswrapper[4948]: I0312 00:08:22.706614 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:22 crc kubenswrapper[4948]: I0312 00:08:22.706685 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:22 crc kubenswrapper[4948]: I0312 00:08:22.706708 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:22 crc kubenswrapper[4948]: I0312 00:08:22.706736 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:22 crc kubenswrapper[4948]: I0312 00:08:22.706758 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:22Z","lastTransitionTime":"2026-03-12T00:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:22 crc kubenswrapper[4948]: I0312 00:08:22.810112 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:22 crc kubenswrapper[4948]: I0312 00:08:22.810176 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:22 crc kubenswrapper[4948]: I0312 00:08:22.810192 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:22 crc kubenswrapper[4948]: I0312 00:08:22.810216 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:22 crc kubenswrapper[4948]: I0312 00:08:22.810238 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:22Z","lastTransitionTime":"2026-03-12T00:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:22 crc kubenswrapper[4948]: I0312 00:08:22.913979 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:22 crc kubenswrapper[4948]: I0312 00:08:22.914032 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:22 crc kubenswrapper[4948]: I0312 00:08:22.914054 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:22 crc kubenswrapper[4948]: I0312 00:08:22.914084 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:22 crc kubenswrapper[4948]: I0312 00:08:22.914103 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:22Z","lastTransitionTime":"2026-03-12T00:08:22Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:23 crc kubenswrapper[4948]: I0312 00:08:23.017482 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:23 crc kubenswrapper[4948]: I0312 00:08:23.017553 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:23 crc kubenswrapper[4948]: I0312 00:08:23.017569 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:23 crc kubenswrapper[4948]: I0312 00:08:23.017592 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:23 crc kubenswrapper[4948]: I0312 00:08:23.017615 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:23Z","lastTransitionTime":"2026-03-12T00:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:23 crc kubenswrapper[4948]: I0312 00:08:23.119851 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:23 crc kubenswrapper[4948]: I0312 00:08:23.120070 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:23 crc kubenswrapper[4948]: I0312 00:08:23.120173 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:23 crc kubenswrapper[4948]: I0312 00:08:23.120264 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:23 crc kubenswrapper[4948]: I0312 00:08:23.120397 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:23Z","lastTransitionTime":"2026-03-12T00:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:23 crc kubenswrapper[4948]: I0312 00:08:23.223105 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:23 crc kubenswrapper[4948]: I0312 00:08:23.223150 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:23 crc kubenswrapper[4948]: I0312 00:08:23.223158 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:23 crc kubenswrapper[4948]: I0312 00:08:23.223172 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:23 crc kubenswrapper[4948]: I0312 00:08:23.223185 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:23Z","lastTransitionTime":"2026-03-12T00:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:23 crc kubenswrapper[4948]: I0312 00:08:23.326071 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:23 crc kubenswrapper[4948]: I0312 00:08:23.326113 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:23 crc kubenswrapper[4948]: I0312 00:08:23.326121 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:23 crc kubenswrapper[4948]: I0312 00:08:23.326174 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:23 crc kubenswrapper[4948]: I0312 00:08:23.326184 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:23Z","lastTransitionTime":"2026-03-12T00:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:23 crc kubenswrapper[4948]: I0312 00:08:23.429956 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:23 crc kubenswrapper[4948]: I0312 00:08:23.430259 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:23 crc kubenswrapper[4948]: I0312 00:08:23.430465 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:23 crc kubenswrapper[4948]: I0312 00:08:23.430678 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:23 crc kubenswrapper[4948]: I0312 00:08:23.430833 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:23Z","lastTransitionTime":"2026-03-12T00:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:23 crc kubenswrapper[4948]: I0312 00:08:23.535649 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:23 crc kubenswrapper[4948]: I0312 00:08:23.535687 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:23 crc kubenswrapper[4948]: I0312 00:08:23.535701 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:23 crc kubenswrapper[4948]: I0312 00:08:23.535720 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:23 crc kubenswrapper[4948]: I0312 00:08:23.535736 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:23Z","lastTransitionTime":"2026-03-12T00:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:23 crc kubenswrapper[4948]: I0312 00:08:23.638548 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:23 crc kubenswrapper[4948]: I0312 00:08:23.638614 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:23 crc kubenswrapper[4948]: I0312 00:08:23.638636 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:23 crc kubenswrapper[4948]: I0312 00:08:23.638666 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:23 crc kubenswrapper[4948]: I0312 00:08:23.638686 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:23Z","lastTransitionTime":"2026-03-12T00:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:23 crc kubenswrapper[4948]: I0312 00:08:23.740769 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:23 crc kubenswrapper[4948]: I0312 00:08:23.740805 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:23 crc kubenswrapper[4948]: I0312 00:08:23.740815 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:23 crc kubenswrapper[4948]: I0312 00:08:23.740831 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:23 crc kubenswrapper[4948]: I0312 00:08:23.740844 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:23Z","lastTransitionTime":"2026-03-12T00:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:23 crc kubenswrapper[4948]: I0312 00:08:23.843863 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:23 crc kubenswrapper[4948]: I0312 00:08:23.843957 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:23 crc kubenswrapper[4948]: I0312 00:08:23.843992 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:23 crc kubenswrapper[4948]: I0312 00:08:23.844024 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:23 crc kubenswrapper[4948]: I0312 00:08:23.844048 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:23Z","lastTransitionTime":"2026-03-12T00:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:23 crc kubenswrapper[4948]: I0312 00:08:23.947142 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:23 crc kubenswrapper[4948]: I0312 00:08:23.947189 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:23 crc kubenswrapper[4948]: I0312 00:08:23.947203 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:23 crc kubenswrapper[4948]: I0312 00:08:23.947223 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:23 crc kubenswrapper[4948]: I0312 00:08:23.947240 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:23Z","lastTransitionTime":"2026-03-12T00:08:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.050938 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.050992 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.051008 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.051032 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.051051 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:24Z","lastTransitionTime":"2026-03-12T00:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.158431 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.159220 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.159245 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.159275 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.159295 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:24Z","lastTransitionTime":"2026-03-12T00:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.262009 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.262120 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.262226 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.262252 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.262355 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:24Z","lastTransitionTime":"2026-03-12T00:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.311596 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.311634 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.311696 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.311706 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:08:24 crc kubenswrapper[4948]: E0312 00:08:24.311769 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 00:08:24 crc kubenswrapper[4948]: E0312 00:08:24.311829 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mqmtd" podUID="be985ff1-de8a-431f-a230-67894ebe52d5" Mar 12 00:08:24 crc kubenswrapper[4948]: E0312 00:08:24.311934 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 00:08:24 crc kubenswrapper[4948]: E0312 00:08:24.312064 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.366034 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.366116 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.366133 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.366158 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.366176 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:24Z","lastTransitionTime":"2026-03-12T00:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.468860 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.468931 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.468948 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.468976 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.468993 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:24Z","lastTransitionTime":"2026-03-12T00:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.572097 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.572194 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.572217 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.572241 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.572259 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:24Z","lastTransitionTime":"2026-03-12T00:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.648661 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f7mp5_15117613-9fad-48c7-98c4-a2d84502ded9/ovnkube-controller/0.log" Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.653657 4948 generic.go:334] "Generic (PLEG): container finished" podID="15117613-9fad-48c7-98c4-a2d84502ded9" containerID="1610d74076d215d3451c6503afb264abd468647b65c5da3e2cf8983ed1aeaeb2" exitCode=1 Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.653719 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" event={"ID":"15117613-9fad-48c7-98c4-a2d84502ded9","Type":"ContainerDied","Data":"1610d74076d215d3451c6503afb264abd468647b65c5da3e2cf8983ed1aeaeb2"} Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.654838 4948 scope.go:117] "RemoveContainer" containerID="1610d74076d215d3451c6503afb264abd468647b65c5da3e2cf8983ed1aeaeb2" Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.675601 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.675660 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.675682 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.675709 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.675729 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:24Z","lastTransitionTime":"2026-03-12T00:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.686434 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15117613-9fad-48c7-98c4-a2d84502ded9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4036d7131db00afd8eab8e093626b688297dc776b5af80c3c38b51da7124085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84af4cdd213d90636a889a8a8a3aaca3dee1e2677368efa0ffc768940c2cc994\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e481ace9e84b6aa66cbdda3ee8425c4120a665915d2c70ae4faf60bba6205e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9eadefd0babc119f776d48c9d2c9e4bced58db685af9a8e97f25e6a9a2e3c89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9d5e77753fda70350353daa0967257dd4929dba4440a95495ab4c2a0659aef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cad943edb9a8d445b33a6e70d85b475c1ccba460a6056bccae94fe60fd6b73f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1610d74076d215d3451c6503afb264abd468647b65c5da3e2cf8983ed1aeaeb2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1610d74076d215d3451c6503afb264abd468647b65c5da3e2cf8983ed1aeaeb2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T00:08:23Z\\\",\\\"message\\\":\\\" 6793 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0312 00:08:23.923364 6793 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0312 00:08:23.923474 6793 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0312 00:08:23.923542 6793 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0312 00:08:23.923552 6793 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0312 00:08:23.923635 6793 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0312 00:08:23.923647 6793 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0312 00:08:23.923766 6793 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0312 00:08:23.923844 6793 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0312 00:08:23.923915 6793 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0312 00:08:23.923976 6793 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0312 00:08:23.924040 6793 handler.go:208] Removed *v1.Node event handler 2\\\\nI0312 00:08:23.924050 6793 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0312 00:08:23.924092 6793 factory.go:656] Stopping watch factory\\\\nI0312 00:08:23.924107 6793 handler.go:208] Removed *v1.Node event handler 7\\\\nI0312 00:08:23.924116 6793 ovnkube.go:599] Stopped ovnkube\\\\nI0312 00:08:2\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:21Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46fa6140f4f4bd32e298fa36f6d7af6890029f9bc3080d2b4d94be8307346de8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7mp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:24Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.706990 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:24Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.725326 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:24Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.744466 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8d58ee45d75284e030700fa526b2c39197a26f6a391a7417f25a026f356b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:24Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.757417 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mqmtd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be985ff1-de8a-431f-a230-67894ebe52d5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmsz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmsz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mqmtd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:24Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.778512 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.778570 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.778589 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.778618 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.778636 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:24Z","lastTransitionTime":"2026-03-12T00:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.779957 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfp2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"662dcbe7-7cf8-42d4-9a5e-345c32a49972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff3f3c69dcb28c7b0fd85f9e0c33b3339a1f6159dbf10bcf0efbdce4f7728c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a57d476b5b00110168878bd1c19c32ef122525c0259ac0a70de31c8220996d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4a57d476b5b00110168878bd1c19c32ef122525c0259ac0a70de31c8220996d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bb429b94a1286d8aac77cb77112c7163f7aa50a1c9cab997f8d088c100887c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0bb429b94a1286d8aac77cb77112c7163f7aa50a1c9cab997f8d088c100887c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a77485fd9e246f190724b86663e0dbe81146e1fcc51f468dd05f23c1db878fd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a77485fd9e246f190724b86663e0dbe81146e1fcc51f468dd05f23c1db878fd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adaa20d9a4898057802a8c9759ab3cd485963180bfc4f6b926bf6d2c886bcf27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adaa20d9a4898057802a8c9759ab3cd485963180bfc4f6b926bf6d2c886bcf27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://998fbb934f01c9ecd315f78e776199d4fd538106f8373f09b730d2df23ac9e3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://998fbb934f01c9ecd315f78e776199d4fd538106f8373f09b730d2df23ac9e3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46ebe18804ff07eb0e922a6cc1c39ff93875302d231222642b211e3527fe6373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46ebe18804ff07eb0e922a6cc1c39ff93875302d231222642b211e3527fe6373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfp2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:24Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.801575 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e483c2b-08f0-4e92-8e4a-b7281f30af3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d6839c99ab958ad1f40636c70e29bc285292fdbee1e910cdf7f8ab43ca7e8db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd9jv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82de8f2b633f1b6c295149d077bfd2b308f5acaa429acf411ed16f456536d0ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd9jv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4xwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:24Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.820370 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5139686878eb07d3abced92d0ccfe05be8907580ed75233b72fc2ad3415a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15cbe3fa75b69a7176803295e16a314a8bc73f5797c874124d54e76d47fa1064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:24Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.833742 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:24Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.847630 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-njc9c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0d6440c-4524-42ea-b18a-0ddc4662f672\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f4d0d9599ad62aaba35ba1414a806f3a163e29a3c34fd893ef5b14eae003591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7wn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-njc9c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:24Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.863783 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2efff53-8e12-41a1-bcb1-9a1d08c42a4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1cb328c8431b39dac735d49726a2eae40a367849b2c146bb2e49d02958f0e5a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ad13662849131222d5c8ba1041b81fa75bcb1fcbfa2e71c0c751f04b532a5b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8cd4faf47b25395ffd42d4bf5416ea29dcaa3ec380d6d6d6a9f13321d35195a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73834c8db23f6536034dce09e38b56665211238280bf9a4e1b13e74d81245a77\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73834c8db23f6536034dce09e38b56665211238280bf9a4e1b13e74d81245a77\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T00:07:50Z\\\",\\\"message\\\":\\\"le observer\\\\nW0312 00:07:49.927227 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0312 00:07:49.927347 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 00:07:49.928000 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2511358960/tls.crt::/tmp/serving-cert-2511358960/tls.key\\\\\\\"\\\\nI0312 00:07:50.321455 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0312 00:07:50.324479 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0312 00:07:50.324511 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0312 00:07:50.324546 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0312 00:07:50.324556 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0312 00:07:50.331591 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0312 00:07:50.331609 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0312 00:07:50.331634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 00:07:50.331642 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 00:07:50.331651 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0312 00:07:50.331658 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0312 00:07:50.331664 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0312 00:07:50.331670 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0312 00:07:50.332917 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:07:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d56e32f9cb96ad103bab80d6a730b70d3387264a25a7a61ef1f0726e7290d6c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e42f2e94463ac8a3d7980a25122062da9ee87fcdc1f10e260ff9974b2e42359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e42f2e94463ac8a3d7980a25122062da9ee87fcdc1f10e260ff9974b2e42359\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:06:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:24Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.876777 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5a9ae9441886817563a5fc793bff7c9b8d77d406ceb1d1fb1eebd4307a1c64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:24Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.880506 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.880540 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.880553 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.880571 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.880584 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:24Z","lastTransitionTime":"2026-03-12T00:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.895476 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwpqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b7580c2-5c23-4c67-807a-ea97a3df9398\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d07c603619f858177d4b007e785208853abcfe2dc6c1af435309030fa409010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8lxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwpqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:24Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.906527 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9s5fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f9c7f1-6593-4d92-854c-e6c9964a0435\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b8833cbaf3809463e18bfebc9d5bada5fb26823831cc742a3a27177662f60e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xf4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9s5fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:24Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.918842 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xdd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b5233d6-d2b8-42dc-b881-18e439ecc442\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0ea55d993c72b6d50a4829ecfff4adae7d0f625a90e40545c9e63201cdda29c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjpwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb23ff4bc484524e306561c0c574e2b2eee39fc80efcd9f44f1924fe42df1a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjpwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4xdd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:24Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.983381 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.983434 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.983452 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.983477 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:24 crc kubenswrapper[4948]: I0312 00:08:24.983494 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:24Z","lastTransitionTime":"2026-03-12T00:08:24Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:25 crc kubenswrapper[4948]: I0312 00:08:25.086617 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:25 crc kubenswrapper[4948]: I0312 00:08:25.086691 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:25 crc kubenswrapper[4948]: I0312 00:08:25.086711 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:25 crc kubenswrapper[4948]: I0312 00:08:25.086734 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:25 crc kubenswrapper[4948]: I0312 00:08:25.086750 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:25Z","lastTransitionTime":"2026-03-12T00:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:25 crc kubenswrapper[4948]: I0312 00:08:25.190032 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:25 crc kubenswrapper[4948]: I0312 00:08:25.190107 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:25 crc kubenswrapper[4948]: I0312 00:08:25.190130 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:25 crc kubenswrapper[4948]: I0312 00:08:25.190179 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:25 crc kubenswrapper[4948]: I0312 00:08:25.190204 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:25Z","lastTransitionTime":"2026-03-12T00:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:25 crc kubenswrapper[4948]: I0312 00:08:25.292394 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:25 crc kubenswrapper[4948]: I0312 00:08:25.292439 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:25 crc kubenswrapper[4948]: I0312 00:08:25.292450 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:25 crc kubenswrapper[4948]: I0312 00:08:25.292466 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:25 crc kubenswrapper[4948]: I0312 00:08:25.292478 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:25Z","lastTransitionTime":"2026-03-12T00:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:25 crc kubenswrapper[4948]: I0312 00:08:25.395344 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:25 crc kubenswrapper[4948]: I0312 00:08:25.395411 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:25 crc kubenswrapper[4948]: I0312 00:08:25.395435 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:25 crc kubenswrapper[4948]: I0312 00:08:25.395465 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:25 crc kubenswrapper[4948]: I0312 00:08:25.395483 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:25Z","lastTransitionTime":"2026-03-12T00:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:25 crc kubenswrapper[4948]: I0312 00:08:25.497813 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:25 crc kubenswrapper[4948]: I0312 00:08:25.497872 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:25 crc kubenswrapper[4948]: I0312 00:08:25.497888 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:25 crc kubenswrapper[4948]: I0312 00:08:25.497911 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:25 crc kubenswrapper[4948]: I0312 00:08:25.497931 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:25Z","lastTransitionTime":"2026-03-12T00:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:25 crc kubenswrapper[4948]: I0312 00:08:25.601175 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:25 crc kubenswrapper[4948]: I0312 00:08:25.601236 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:25 crc kubenswrapper[4948]: I0312 00:08:25.601253 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:25 crc kubenswrapper[4948]: I0312 00:08:25.601277 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:25 crc kubenswrapper[4948]: I0312 00:08:25.601296 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:25Z","lastTransitionTime":"2026-03-12T00:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:25 crc kubenswrapper[4948]: I0312 00:08:25.659276 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f7mp5_15117613-9fad-48c7-98c4-a2d84502ded9/ovnkube-controller/0.log" Mar 12 00:08:25 crc kubenswrapper[4948]: I0312 00:08:25.662843 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" event={"ID":"15117613-9fad-48c7-98c4-a2d84502ded9","Type":"ContainerStarted","Data":"588e8a26bafbb970547438b573ab34f111ef18d8134df2b3fff513c9cca9a516"} Mar 12 00:08:25 crc kubenswrapper[4948]: I0312 00:08:25.663419 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:25 crc kubenswrapper[4948]: I0312 00:08:25.688746 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2efff53-8e12-41a1-bcb1-9a1d08c42a4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1cb328c8431b39dac735d49726a2eae40a367849b2c146bb2e49d02958f0e5a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ad13662849131222d5c8ba1041b81fa75bcb1fcbfa2e71c0c751f04b532a5b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8cd4faf47b25395ffd42d4bf5416ea29dcaa3ec380d6d6d6a9f13321d35195a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73834c8db23f6536034dce09e38b56665211238280bf9a4e1b13e74d81245a77\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73834c8db23f6536034dce09e38b56665211238280bf9a4e1b13e74d81245a77\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T00:07:50Z\\\",\\\"message\\\":\\\"le observer\\\\nW0312 00:07:49.927227 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0312 00:07:49.927347 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 00:07:49.928000 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2511358960/tls.crt::/tmp/serving-cert-2511358960/tls.key\\\\\\\"\\\\nI0312 00:07:50.321455 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0312 00:07:50.324479 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0312 00:07:50.324511 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0312 00:07:50.324546 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0312 00:07:50.324556 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0312 00:07:50.331591 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0312 00:07:50.331609 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0312 00:07:50.331634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 00:07:50.331642 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 00:07:50.331651 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0312 00:07:50.331658 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0312 00:07:50.331664 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0312 00:07:50.331670 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0312 00:07:50.332917 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:07:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d56e32f9cb96ad103bab80d6a730b70d3387264a25a7a61ef1f0726e7290d6c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e42f2e94463ac8a3d7980a25122062da9ee87fcdc1f10e260ff9974b2e42359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e42f2e94463ac8a3d7980a25122062da9ee87fcdc1f10e260ff9974b2e42359\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:06:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:25Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:25 crc kubenswrapper[4948]: I0312 00:08:25.704240 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:25 crc kubenswrapper[4948]: I0312 00:08:25.704356 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:25 crc kubenswrapper[4948]: I0312 00:08:25.704382 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:25 crc kubenswrapper[4948]: I0312 00:08:25.704412 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:25 crc kubenswrapper[4948]: I0312 00:08:25.704438 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:25Z","lastTransitionTime":"2026-03-12T00:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:25 crc kubenswrapper[4948]: I0312 00:08:25.711402 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5a9ae9441886817563a5fc793bff7c9b8d77d406ceb1d1fb1eebd4307a1c64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:25Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:25 crc kubenswrapper[4948]: I0312 00:08:25.737481 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwpqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b7580c2-5c23-4c67-807a-ea97a3df9398\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d07c603619f858177d4b007e785208853abcfe2dc6c1af435309030fa409010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8lxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwpqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:25Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:25 crc kubenswrapper[4948]: I0312 00:08:25.751123 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-njc9c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0d6440c-4524-42ea-b18a-0ddc4662f672\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f4d0d9599ad62aaba35ba1414a806f3a163e29a3c34fd893ef5b14eae003591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7wn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-njc9c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:25Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:25 crc kubenswrapper[4948]: I0312 00:08:25.766972 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9s5fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f9c7f1-6593-4d92-854c-e6c9964a0435\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b8833cbaf3809463e18bfebc9d5bada5fb26823831cc742a3a27177662f60e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xf4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9s5fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:25Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:25 crc kubenswrapper[4948]: I0312 00:08:25.777922 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xdd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b5233d6-d2b8-42dc-b881-18e439ecc442\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0ea55d993c72b6d50a4829ecfff4adae7d0f625a90e40545c9e63201cdda29c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjpwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb23ff4bc484524e306561c0c574e2b2eee39fc80efcd9f44f1924fe42df1a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjpwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4xdd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:25Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:25 crc kubenswrapper[4948]: I0312 00:08:25.792615 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:25Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:25 crc kubenswrapper[4948]: I0312 00:08:25.807049 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:25 crc kubenswrapper[4948]: I0312 00:08:25.807127 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:25 crc kubenswrapper[4948]: I0312 00:08:25.807141 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:25 crc kubenswrapper[4948]: I0312 00:08:25.807158 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:25 crc kubenswrapper[4948]: I0312 00:08:25.807171 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:25Z","lastTransitionTime":"2026-03-12T00:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:25 crc kubenswrapper[4948]: I0312 00:08:25.809492 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:25Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:25 crc kubenswrapper[4948]: I0312 00:08:25.822457 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8d58ee45d75284e030700fa526b2c39197a26f6a391a7417f25a026f356b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:25Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:25 crc kubenswrapper[4948]: I0312 00:08:25.835527 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mqmtd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be985ff1-de8a-431f-a230-67894ebe52d5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmsz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmsz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mqmtd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:25Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:25 crc kubenswrapper[4948]: I0312 00:08:25.859278 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15117613-9fad-48c7-98c4-a2d84502ded9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4036d7131db00afd8eab8e093626b688297dc776b5af80c3c38b51da7124085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84af4cdd213d90636a889a8a8a3aaca3dee1e2677368efa0ffc768940c2cc994\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e481ace9e84b6aa66cbdda3ee8425c4120a665915d2c70ae4faf60bba6205e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9eadefd0babc119f776d48c9d2c9e4bced58db685af9a8e97f25e6a9a2e3c89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9d5e77753fda70350353daa0967257dd4929dba4440a95495ab4c2a0659aef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cad943edb9a8d445b33a6e70d85b475c1ccba460a6056bccae94fe60fd6b73f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://588e8a26bafbb970547438b573ab34f111ef18d8134df2b3fff513c9cca9a516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1610d74076d215d3451c6503afb264abd468647b65c5da3e2cf8983ed1aeaeb2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T00:08:23Z\\\",\\\"message\\\":\\\" 6793 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0312 00:08:23.923364 6793 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0312 00:08:23.923474 6793 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0312 00:08:23.923542 6793 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0312 00:08:23.923552 6793 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0312 00:08:23.923635 6793 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0312 00:08:23.923647 6793 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0312 00:08:23.923766 6793 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0312 00:08:23.923844 6793 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0312 00:08:23.923915 6793 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0312 00:08:23.923976 6793 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0312 00:08:23.924040 6793 handler.go:208] Removed *v1.Node event handler 2\\\\nI0312 00:08:23.924050 6793 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0312 00:08:23.924092 6793 factory.go:656] Stopping watch factory\\\\nI0312 00:08:23.924107 6793 handler.go:208] Removed *v1.Node event handler 7\\\\nI0312 00:08:23.924116 6793 ovnkube.go:599] Stopped ovnkube\\\\nI0312 00:08:2\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46fa6140f4f4bd32e298fa36f6d7af6890029f9bc3080d2b4d94be8307346de8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7mp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:25Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:25 crc kubenswrapper[4948]: I0312 00:08:25.870612 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e483c2b-08f0-4e92-8e4a-b7281f30af3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d6839c99ab958ad1f40636c70e29bc285292fdbee1e910cdf7f8ab43ca7e8db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd9jv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82de8f2b633f1b6c295149d077bfd2b308f5acaa429acf411ed16f456536d0ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd9jv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4xwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:25Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:25 crc kubenswrapper[4948]: I0312 00:08:25.889956 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5139686878eb07d3abced92d0ccfe05be8907580ed75233b72fc2ad3415a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15cbe3fa75b69a7176803295e16a314a8bc73f5797c874124d54e76d47fa1064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:25Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:25 crc kubenswrapper[4948]: I0312 00:08:25.909913 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:25 crc kubenswrapper[4948]: I0312 00:08:25.909973 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:25 crc kubenswrapper[4948]: I0312 00:08:25.909994 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:25 crc kubenswrapper[4948]: I0312 00:08:25.910018 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:25 crc kubenswrapper[4948]: I0312 00:08:25.910038 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:25Z","lastTransitionTime":"2026-03-12T00:08:25Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:25 crc kubenswrapper[4948]: I0312 00:08:25.910777 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:25Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:25 crc kubenswrapper[4948]: I0312 00:08:25.937679 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfp2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"662dcbe7-7cf8-42d4-9a5e-345c32a49972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff3f3c69dcb28c7b0fd85f9e0c33b3339a1f6159dbf10bcf0efbdce4f7728c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a57d476b5b00110168878bd1c19c32ef122525c0259ac0a70de31c8220996d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4a57d476b5b00110168878bd1c19c32ef122525c0259ac0a70de31c8220996d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bb429b94a1286d8aac77cb77112c7163f7aa50a1c9cab997f8d088c100887c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0bb429b94a1286d8aac77cb77112c7163f7aa50a1c9cab997f8d088c100887c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a77485fd9e246f190724b86663e0dbe81146e1fcc51f468dd05f23c1db878fd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a77485fd9e246f190724b86663e0dbe81146e1fcc51f468dd05f23c1db878fd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adaa20d9a4898057802a8c9759ab3cd485963180bfc4f6b926bf6d2c886bcf27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adaa20d9a4898057802a8c9759ab3cd485963180bfc4f6b926bf6d2c886bcf27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://998fbb934f01c9ecd315f78e776199d4fd538106f8373f09b730d2df23ac9e3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://998fbb934f01c9ecd315f78e776199d4fd538106f8373f09b730d2df23ac9e3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46ebe18804ff07eb0e922a6cc1c39ff93875302d231222642b211e3527fe6373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46ebe18804ff07eb0e922a6cc1c39ff93875302d231222642b211e3527fe6373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfp2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:25Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.012809 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.012882 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.012900 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.012928 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.012945 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:26Z","lastTransitionTime":"2026-03-12T00:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.115556 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.115607 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.115626 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.115651 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.115669 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:26Z","lastTransitionTime":"2026-03-12T00:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.219052 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.219098 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.219115 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.219139 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.219157 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:26Z","lastTransitionTime":"2026-03-12T00:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.312206 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.312369 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:08:26 crc kubenswrapper[4948]: E0312 00:08:26.312562 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.312592 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.312652 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:08:26 crc kubenswrapper[4948]: E0312 00:08:26.312792 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 00:08:26 crc kubenswrapper[4948]: E0312 00:08:26.312957 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 00:08:26 crc kubenswrapper[4948]: E0312 00:08:26.313119 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mqmtd" podUID="be985ff1-de8a-431f-a230-67894ebe52d5" Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.322007 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.322055 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.322066 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.322082 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.322094 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:26Z","lastTransitionTime":"2026-03-12T00:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.424766 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.424821 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.424837 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.424882 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.424900 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:26Z","lastTransitionTime":"2026-03-12T00:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.527557 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.527615 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.527631 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.527652 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.527667 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:26Z","lastTransitionTime":"2026-03-12T00:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.630864 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.630930 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.630958 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.630985 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.631009 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:26Z","lastTransitionTime":"2026-03-12T00:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.668573 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f7mp5_15117613-9fad-48c7-98c4-a2d84502ded9/ovnkube-controller/1.log" Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.669467 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f7mp5_15117613-9fad-48c7-98c4-a2d84502ded9/ovnkube-controller/0.log" Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.673043 4948 generic.go:334] "Generic (PLEG): container finished" podID="15117613-9fad-48c7-98c4-a2d84502ded9" containerID="588e8a26bafbb970547438b573ab34f111ef18d8134df2b3fff513c9cca9a516" exitCode=1 Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.673087 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" event={"ID":"15117613-9fad-48c7-98c4-a2d84502ded9","Type":"ContainerDied","Data":"588e8a26bafbb970547438b573ab34f111ef18d8134df2b3fff513c9cca9a516"} Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.673123 4948 scope.go:117] "RemoveContainer" containerID="1610d74076d215d3451c6503afb264abd468647b65c5da3e2cf8983ed1aeaeb2" Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.673945 4948 scope.go:117] "RemoveContainer" containerID="588e8a26bafbb970547438b573ab34f111ef18d8134df2b3fff513c9cca9a516" Mar 12 00:08:26 crc kubenswrapper[4948]: E0312 00:08:26.674135 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-f7mp5_openshift-ovn-kubernetes(15117613-9fad-48c7-98c4-a2d84502ded9)\"" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" podUID="15117613-9fad-48c7-98c4-a2d84502ded9" Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.696652 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5139686878eb07d3abced92d0ccfe05be8907580ed75233b72fc2ad3415a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15cbe3fa75b69a7176803295e16a314a8bc73f5797c874124d54e76d47fa1064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:26Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.715721 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:26Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.733243 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.733421 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.733504 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.733597 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.733685 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:26Z","lastTransitionTime":"2026-03-12T00:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.739041 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfp2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"662dcbe7-7cf8-42d4-9a5e-345c32a49972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff3f3c69dcb28c7b0fd85f9e0c33b3339a1f6159dbf10bcf0efbdce4f7728c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a57d476b5b00110168878bd1c19c32ef122525c0259ac0a70de31c8220996d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4a57d476b5b00110168878bd1c19c32ef122525c0259ac0a70de31c8220996d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bb429b94a1286d8aac77cb77112c7163f7aa50a1c9cab997f8d088c100887c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0bb429b94a1286d8aac77cb77112c7163f7aa50a1c9cab997f8d088c100887c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a77485fd9e246f190724b86663e0dbe81146e1fcc51f468dd05f23c1db878fd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a77485fd9e246f190724b86663e0dbe81146e1fcc51f468dd05f23c1db878fd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adaa20d9a4898057802a8c9759ab3cd485963180bfc4f6b926bf6d2c886bcf27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adaa20d9a4898057802a8c9759ab3cd485963180bfc4f6b926bf6d2c886bcf27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://998fbb934f01c9ecd315f78e776199d4fd538106f8373f09b730d2df23ac9e3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://998fbb934f01c9ecd315f78e776199d4fd538106f8373f09b730d2df23ac9e3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46ebe18804ff07eb0e922a6cc1c39ff93875302d231222642b211e3527fe6373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46ebe18804ff07eb0e922a6cc1c39ff93875302d231222642b211e3527fe6373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfp2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:26Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.755100 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e483c2b-08f0-4e92-8e4a-b7281f30af3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d6839c99ab958ad1f40636c70e29bc285292fdbee1e910cdf7f8ab43ca7e8db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd9jv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82de8f2b633f1b6c295149d077bfd2b308f5acaa429acf411ed16f456536d0ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd9jv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4xwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:26Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.810001 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2efff53-8e12-41a1-bcb1-9a1d08c42a4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1cb328c8431b39dac735d49726a2eae40a367849b2c146bb2e49d02958f0e5a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ad13662849131222d5c8ba1041b81fa75bcb1fcbfa2e71c0c751f04b532a5b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8cd4faf47b25395ffd42d4bf5416ea29dcaa3ec380d6d6d6a9f13321d35195a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73834c8db23f6536034dce09e38b56665211238280bf9a4e1b13e74d81245a77\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73834c8db23f6536034dce09e38b56665211238280bf9a4e1b13e74d81245a77\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T00:07:50Z\\\",\\\"message\\\":\\\"le observer\\\\nW0312 00:07:49.927227 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0312 00:07:49.927347 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 00:07:49.928000 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2511358960/tls.crt::/tmp/serving-cert-2511358960/tls.key\\\\\\\"\\\\nI0312 00:07:50.321455 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0312 00:07:50.324479 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0312 00:07:50.324511 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0312 00:07:50.324546 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0312 00:07:50.324556 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0312 00:07:50.331591 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0312 00:07:50.331609 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0312 00:07:50.331634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 00:07:50.331642 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 00:07:50.331651 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0312 00:07:50.331658 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0312 00:07:50.331664 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0312 00:07:50.331670 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0312 00:07:50.332917 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:07:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d56e32f9cb96ad103bab80d6a730b70d3387264a25a7a61ef1f0726e7290d6c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e42f2e94463ac8a3d7980a25122062da9ee87fcdc1f10e260ff9974b2e42359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e42f2e94463ac8a3d7980a25122062da9ee87fcdc1f10e260ff9974b2e42359\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:06:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:26Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.825361 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5a9ae9441886817563a5fc793bff7c9b8d77d406ceb1d1fb1eebd4307a1c64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:26Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.836142 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.836192 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.836209 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.836230 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.836246 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:26Z","lastTransitionTime":"2026-03-12T00:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.841947 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwpqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b7580c2-5c23-4c67-807a-ea97a3df9398\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d07c603619f858177d4b007e785208853abcfe2dc6c1af435309030fa409010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8lxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwpqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:26Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.856928 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-njc9c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0d6440c-4524-42ea-b18a-0ddc4662f672\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f4d0d9599ad62aaba35ba1414a806f3a163e29a3c34fd893ef5b14eae003591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7wn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-njc9c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:26Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.870522 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9s5fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f9c7f1-6593-4d92-854c-e6c9964a0435\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b8833cbaf3809463e18bfebc9d5bada5fb26823831cc742a3a27177662f60e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xf4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9s5fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:26Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.883835 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xdd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b5233d6-d2b8-42dc-b881-18e439ecc442\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0ea55d993c72b6d50a4829ecfff4adae7d0f625a90e40545c9e63201cdda29c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjpwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb23ff4bc484524e306561c0c574e2b2eee39fc80efcd9f44f1924fe42df1a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjpwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4xdd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:26Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.898492 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:26Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.915242 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:26Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.934770 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8d58ee45d75284e030700fa526b2c39197a26f6a391a7417f25a026f356b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:26Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.938170 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.938231 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.938246 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.938269 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.938284 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:26Z","lastTransitionTime":"2026-03-12T00:08:26Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.945986 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mqmtd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be985ff1-de8a-431f-a230-67894ebe52d5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmsz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmsz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mqmtd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:26Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:26 crc kubenswrapper[4948]: I0312 00:08:26.974078 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15117613-9fad-48c7-98c4-a2d84502ded9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4036d7131db00afd8eab8e093626b688297dc776b5af80c3c38b51da7124085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84af4cdd213d90636a889a8a8a3aaca3dee1e2677368efa0ffc768940c2cc994\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e481ace9e84b6aa66cbdda3ee8425c4120a665915d2c70ae4faf60bba6205e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9eadefd0babc119f776d48c9d2c9e4bced58db685af9a8e97f25e6a9a2e3c89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9d5e77753fda70350353daa0967257dd4929dba4440a95495ab4c2a0659aef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cad943edb9a8d445b33a6e70d85b475c1ccba460a6056bccae94fe60fd6b73f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://588e8a26bafbb970547438b573ab34f111ef18d8134df2b3fff513c9cca9a516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1610d74076d215d3451c6503afb264abd468647b65c5da3e2cf8983ed1aeaeb2\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T00:08:23Z\\\",\\\"message\\\":\\\" 6793 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0312 00:08:23.923364 6793 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0312 00:08:23.923474 6793 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0312 00:08:23.923542 6793 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0312 00:08:23.923552 6793 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0312 00:08:23.923635 6793 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0312 00:08:23.923647 6793 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0312 00:08:23.923766 6793 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0312 00:08:23.923844 6793 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0312 00:08:23.923915 6793 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0312 00:08:23.923976 6793 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0312 00:08:23.924040 6793 handler.go:208] Removed *v1.Node event handler 2\\\\nI0312 00:08:23.924050 6793 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0312 00:08:23.924092 6793 factory.go:656] Stopping watch factory\\\\nI0312 00:08:23.924107 6793 handler.go:208] Removed *v1.Node event handler 7\\\\nI0312 00:08:23.924116 6793 ovnkube.go:599] Stopped ovnkube\\\\nI0312 00:08:2\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:21Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://588e8a26bafbb970547438b573ab34f111ef18d8134df2b3fff513c9cca9a516\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T00:08:25Z\\\",\\\"message\\\":\\\"umn _uuid == {f6d604c1-9711-4e25-be6c-79ec28bbad1b}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0312 00:08:25.737942 6942 obj_retry.go:551] Creating *factory.egressNode crc took: 2.187304ms\\\\nI0312 00:08:25.737970 6942 factory.go:1336] Added *v1.Node event handler 7\\\\nI0312 00:08:25.738003 6942 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI0312 00:08:25.738210 6942 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0312 00:08:25.738274 6942 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0312 00:08:25.738325 6942 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0312 00:08:25.738333 6942 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0312 00:08:25.738357 6942 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0312 00:08:25.738373 6942 factory.go:656] Stopping watch factory\\\\nI0312 00:08:25.738386 6942 ovnkube.go:599] Stopped ovnkube\\\\nI0312 00:08:25.738410 6942 handler.go:208] Removed *v1.Node event handler 2\\\\nI0312 00:08:25.738414 6942 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0312 00:08:25.738430 6942 handler.go:208] Removed *v1.Node event handler 7\\\\nI0312 00:08:25.738442 6942 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nF0312 00:08:25.738550 6942 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:24Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46fa6140f4f4bd32e298fa36f6d7af6890029f9bc3080d2b4d94be8307346de8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7mp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:26Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:27 crc kubenswrapper[4948]: I0312 00:08:27.041392 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:27 crc kubenswrapper[4948]: I0312 00:08:27.041460 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:27 crc kubenswrapper[4948]: I0312 00:08:27.041495 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:27 crc kubenswrapper[4948]: I0312 00:08:27.041522 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:27 crc kubenswrapper[4948]: I0312 00:08:27.041542 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:27Z","lastTransitionTime":"2026-03-12T00:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:27 crc kubenswrapper[4948]: I0312 00:08:27.144737 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:27 crc kubenswrapper[4948]: I0312 00:08:27.144828 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:27 crc kubenswrapper[4948]: I0312 00:08:27.144845 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:27 crc kubenswrapper[4948]: I0312 00:08:27.144868 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:27 crc kubenswrapper[4948]: I0312 00:08:27.144886 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:27Z","lastTransitionTime":"2026-03-12T00:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:27 crc kubenswrapper[4948]: I0312 00:08:27.247847 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:27 crc kubenswrapper[4948]: I0312 00:08:27.247902 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:27 crc kubenswrapper[4948]: I0312 00:08:27.247923 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:27 crc kubenswrapper[4948]: I0312 00:08:27.247949 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:27 crc kubenswrapper[4948]: I0312 00:08:27.247965 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:27Z","lastTransitionTime":"2026-03-12T00:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:27 crc kubenswrapper[4948]: I0312 00:08:27.350970 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:27 crc kubenswrapper[4948]: I0312 00:08:27.351030 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:27 crc kubenswrapper[4948]: I0312 00:08:27.351048 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:27 crc kubenswrapper[4948]: I0312 00:08:27.351069 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:27 crc kubenswrapper[4948]: I0312 00:08:27.351105 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:27Z","lastTransitionTime":"2026-03-12T00:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:27 crc kubenswrapper[4948]: I0312 00:08:27.454024 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:27 crc kubenswrapper[4948]: I0312 00:08:27.454085 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:27 crc kubenswrapper[4948]: I0312 00:08:27.454110 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:27 crc kubenswrapper[4948]: I0312 00:08:27.454143 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:27 crc kubenswrapper[4948]: I0312 00:08:27.454164 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:27Z","lastTransitionTime":"2026-03-12T00:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:27 crc kubenswrapper[4948]: I0312 00:08:27.558017 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:27 crc kubenswrapper[4948]: I0312 00:08:27.558075 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:27 crc kubenswrapper[4948]: I0312 00:08:27.558092 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:27 crc kubenswrapper[4948]: I0312 00:08:27.558119 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:27 crc kubenswrapper[4948]: I0312 00:08:27.558138 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:27Z","lastTransitionTime":"2026-03-12T00:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:27 crc kubenswrapper[4948]: I0312 00:08:27.661493 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:27 crc kubenswrapper[4948]: I0312 00:08:27.661571 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:27 crc kubenswrapper[4948]: I0312 00:08:27.661593 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:27 crc kubenswrapper[4948]: I0312 00:08:27.661625 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:27 crc kubenswrapper[4948]: I0312 00:08:27.661648 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:27Z","lastTransitionTime":"2026-03-12T00:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:27 crc kubenswrapper[4948]: I0312 00:08:27.679250 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f7mp5_15117613-9fad-48c7-98c4-a2d84502ded9/ovnkube-controller/1.log" Mar 12 00:08:27 crc kubenswrapper[4948]: I0312 00:08:27.685095 4948 scope.go:117] "RemoveContainer" containerID="588e8a26bafbb970547438b573ab34f111ef18d8134df2b3fff513c9cca9a516" Mar 12 00:08:27 crc kubenswrapper[4948]: E0312 00:08:27.685382 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-f7mp5_openshift-ovn-kubernetes(15117613-9fad-48c7-98c4-a2d84502ded9)\"" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" podUID="15117613-9fad-48c7-98c4-a2d84502ded9" Mar 12 00:08:27 crc kubenswrapper[4948]: I0312 00:08:27.701814 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9s5fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f9c7f1-6593-4d92-854c-e6c9964a0435\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b8833cbaf3809463e18bfebc9d5bada5fb26823831cc742a3a27177662f60e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xf4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9s5fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:27Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:27 crc kubenswrapper[4948]: I0312 00:08:27.719496 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xdd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b5233d6-d2b8-42dc-b881-18e439ecc442\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0ea55d993c72b6d50a4829ecfff4adae7d0f625a90e40545c9e63201cdda29c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjpwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb23ff4bc484524e306561c0c574e2b2eee39fc80efcd9f44f1924fe42df1a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjpwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4xdd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:27Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:27 crc kubenswrapper[4948]: I0312 00:08:27.739473 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:27Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:27 crc kubenswrapper[4948]: I0312 00:08:27.758110 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:27Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:27 crc kubenswrapper[4948]: I0312 00:08:27.764468 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:27 crc kubenswrapper[4948]: I0312 00:08:27.764544 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:27 crc kubenswrapper[4948]: I0312 00:08:27.764569 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:27 crc kubenswrapper[4948]: I0312 00:08:27.764603 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:27 crc kubenswrapper[4948]: I0312 00:08:27.764626 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:27Z","lastTransitionTime":"2026-03-12T00:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:27 crc kubenswrapper[4948]: I0312 00:08:27.777793 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8d58ee45d75284e030700fa526b2c39197a26f6a391a7417f25a026f356b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:27Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:27 crc kubenswrapper[4948]: I0312 00:08:27.794556 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mqmtd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be985ff1-de8a-431f-a230-67894ebe52d5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmsz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmsz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mqmtd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:27Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:27 crc kubenswrapper[4948]: I0312 00:08:27.825268 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15117613-9fad-48c7-98c4-a2d84502ded9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4036d7131db00afd8eab8e093626b688297dc776b5af80c3c38b51da7124085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84af4cdd213d90636a889a8a8a3aaca3dee1e2677368efa0ffc768940c2cc994\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e481ace9e84b6aa66cbdda3ee8425c4120a665915d2c70ae4faf60bba6205e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9eadefd0babc119f776d48c9d2c9e4bced58db685af9a8e97f25e6a9a2e3c89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9d5e77753fda70350353daa0967257dd4929dba4440a95495ab4c2a0659aef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cad943edb9a8d445b33a6e70d85b475c1ccba460a6056bccae94fe60fd6b73f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://588e8a26bafbb970547438b573ab34f111ef18d8134df2b3fff513c9cca9a516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://588e8a26bafbb970547438b573ab34f111ef18d8134df2b3fff513c9cca9a516\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T00:08:25Z\\\",\\\"message\\\":\\\"umn _uuid == {f6d604c1-9711-4e25-be6c-79ec28bbad1b}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0312 00:08:25.737942 6942 obj_retry.go:551] Creating *factory.egressNode crc took: 2.187304ms\\\\nI0312 00:08:25.737970 6942 factory.go:1336] Added *v1.Node event handler 7\\\\nI0312 00:08:25.738003 6942 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI0312 00:08:25.738210 6942 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0312 00:08:25.738274 6942 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0312 00:08:25.738325 6942 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0312 00:08:25.738333 6942 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0312 00:08:25.738357 6942 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0312 00:08:25.738373 6942 factory.go:656] Stopping watch factory\\\\nI0312 00:08:25.738386 6942 ovnkube.go:599] Stopped ovnkube\\\\nI0312 00:08:25.738410 6942 handler.go:208] Removed *v1.Node event handler 2\\\\nI0312 00:08:25.738414 6942 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0312 00:08:25.738430 6942 handler.go:208] Removed *v1.Node event handler 7\\\\nI0312 00:08:25.738442 6942 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nF0312 00:08:25.738550 6942 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:24Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-f7mp5_openshift-ovn-kubernetes(15117613-9fad-48c7-98c4-a2d84502ded9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46fa6140f4f4bd32e298fa36f6d7af6890029f9bc3080d2b4d94be8307346de8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7mp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:27Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:27 crc kubenswrapper[4948]: I0312 00:08:27.848355 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5139686878eb07d3abced92d0ccfe05be8907580ed75233b72fc2ad3415a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15cbe3fa75b69a7176803295e16a314a8bc73f5797c874124d54e76d47fa1064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:27Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:27 crc kubenswrapper[4948]: I0312 00:08:27.867652 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:27 crc kubenswrapper[4948]: I0312 00:08:27.867732 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:27 crc kubenswrapper[4948]: I0312 00:08:27.867753 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:27 crc kubenswrapper[4948]: I0312 00:08:27.867779 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:27 crc kubenswrapper[4948]: I0312 00:08:27.867797 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:27Z","lastTransitionTime":"2026-03-12T00:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:27 crc kubenswrapper[4948]: I0312 00:08:27.868796 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:27Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:27 crc kubenswrapper[4948]: I0312 00:08:27.893211 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfp2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"662dcbe7-7cf8-42d4-9a5e-345c32a49972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff3f3c69dcb28c7b0fd85f9e0c33b3339a1f6159dbf10bcf0efbdce4f7728c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a57d476b5b00110168878bd1c19c32ef122525c0259ac0a70de31c8220996d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4a57d476b5b00110168878bd1c19c32ef122525c0259ac0a70de31c8220996d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bb429b94a1286d8aac77cb77112c7163f7aa50a1c9cab997f8d088c100887c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0bb429b94a1286d8aac77cb77112c7163f7aa50a1c9cab997f8d088c100887c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a77485fd9e246f190724b86663e0dbe81146e1fcc51f468dd05f23c1db878fd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a77485fd9e246f190724b86663e0dbe81146e1fcc51f468dd05f23c1db878fd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adaa20d9a4898057802a8c9759ab3cd485963180bfc4f6b926bf6d2c886bcf27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adaa20d9a4898057802a8c9759ab3cd485963180bfc4f6b926bf6d2c886bcf27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://998fbb934f01c9ecd315f78e776199d4fd538106f8373f09b730d2df23ac9e3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://998fbb934f01c9ecd315f78e776199d4fd538106f8373f09b730d2df23ac9e3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46ebe18804ff07eb0e922a6cc1c39ff93875302d231222642b211e3527fe6373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46ebe18804ff07eb0e922a6cc1c39ff93875302d231222642b211e3527fe6373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfp2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:27Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:27 crc kubenswrapper[4948]: I0312 00:08:27.913216 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e483c2b-08f0-4e92-8e4a-b7281f30af3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d6839c99ab958ad1f40636c70e29bc285292fdbee1e910cdf7f8ab43ca7e8db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd9jv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82de8f2b633f1b6c295149d077bfd2b308f5acaa429acf411ed16f456536d0ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd9jv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4xwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:27Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:27 crc kubenswrapper[4948]: I0312 00:08:27.934553 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2efff53-8e12-41a1-bcb1-9a1d08c42a4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1cb328c8431b39dac735d49726a2eae40a367849b2c146bb2e49d02958f0e5a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ad13662849131222d5c8ba1041b81fa75bcb1fcbfa2e71c0c751f04b532a5b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8cd4faf47b25395ffd42d4bf5416ea29dcaa3ec380d6d6d6a9f13321d35195a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73834c8db23f6536034dce09e38b56665211238280bf9a4e1b13e74d81245a77\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73834c8db23f6536034dce09e38b56665211238280bf9a4e1b13e74d81245a77\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T00:07:50Z\\\",\\\"message\\\":\\\"le observer\\\\nW0312 00:07:49.927227 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0312 00:07:49.927347 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 00:07:49.928000 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2511358960/tls.crt::/tmp/serving-cert-2511358960/tls.key\\\\\\\"\\\\nI0312 00:07:50.321455 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0312 00:07:50.324479 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0312 00:07:50.324511 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0312 00:07:50.324546 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0312 00:07:50.324556 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0312 00:07:50.331591 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0312 00:07:50.331609 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0312 00:07:50.331634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 00:07:50.331642 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 00:07:50.331651 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0312 00:07:50.331658 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0312 00:07:50.331664 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0312 00:07:50.331670 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0312 00:07:50.332917 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:07:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d56e32f9cb96ad103bab80d6a730b70d3387264a25a7a61ef1f0726e7290d6c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e42f2e94463ac8a3d7980a25122062da9ee87fcdc1f10e260ff9974b2e42359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e42f2e94463ac8a3d7980a25122062da9ee87fcdc1f10e260ff9974b2e42359\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:06:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:27Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:27 crc kubenswrapper[4948]: I0312 00:08:27.954231 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5a9ae9441886817563a5fc793bff7c9b8d77d406ceb1d1fb1eebd4307a1c64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:27Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:27 crc kubenswrapper[4948]: I0312 00:08:27.969985 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:27 crc kubenswrapper[4948]: I0312 00:08:27.970045 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:27 crc kubenswrapper[4948]: I0312 00:08:27.970067 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:27 crc kubenswrapper[4948]: I0312 00:08:27.970095 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:27 crc kubenswrapper[4948]: I0312 00:08:27.970117 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:27Z","lastTransitionTime":"2026-03-12T00:08:27Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:27 crc kubenswrapper[4948]: I0312 00:08:27.974147 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwpqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b7580c2-5c23-4c67-807a-ea97a3df9398\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d07c603619f858177d4b007e785208853abcfe2dc6c1af435309030fa409010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8lxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwpqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:27Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:27 crc kubenswrapper[4948]: I0312 00:08:27.988744 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-njc9c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0d6440c-4524-42ea-b18a-0ddc4662f672\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f4d0d9599ad62aaba35ba1414a806f3a163e29a3c34fd893ef5b14eae003591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7wn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-njc9c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:27Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:28 crc kubenswrapper[4948]: I0312 00:08:28.073449 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:28 crc kubenswrapper[4948]: I0312 00:08:28.073509 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:28 crc kubenswrapper[4948]: I0312 00:08:28.073528 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:28 crc kubenswrapper[4948]: I0312 00:08:28.073551 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:28 crc kubenswrapper[4948]: I0312 00:08:28.073570 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:28Z","lastTransitionTime":"2026-03-12T00:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:28 crc kubenswrapper[4948]: I0312 00:08:28.177177 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:28 crc kubenswrapper[4948]: I0312 00:08:28.177234 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:28 crc kubenswrapper[4948]: I0312 00:08:28.177257 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:28 crc kubenswrapper[4948]: I0312 00:08:28.177291 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:28 crc kubenswrapper[4948]: I0312 00:08:28.177350 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:28Z","lastTransitionTime":"2026-03-12T00:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:28 crc kubenswrapper[4948]: I0312 00:08:28.279718 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:28 crc kubenswrapper[4948]: I0312 00:08:28.279901 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:28 crc kubenswrapper[4948]: I0312 00:08:28.279929 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:28 crc kubenswrapper[4948]: I0312 00:08:28.279958 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:28 crc kubenswrapper[4948]: I0312 00:08:28.279979 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:28Z","lastTransitionTime":"2026-03-12T00:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:28 crc kubenswrapper[4948]: I0312 00:08:28.312529 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:08:28 crc kubenswrapper[4948]: I0312 00:08:28.312576 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:08:28 crc kubenswrapper[4948]: I0312 00:08:28.312628 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:08:28 crc kubenswrapper[4948]: I0312 00:08:28.312549 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:08:28 crc kubenswrapper[4948]: E0312 00:08:28.312683 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 00:08:28 crc kubenswrapper[4948]: E0312 00:08:28.312863 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mqmtd" podUID="be985ff1-de8a-431f-a230-67894ebe52d5" Mar 12 00:08:28 crc kubenswrapper[4948]: E0312 00:08:28.312929 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 00:08:28 crc kubenswrapper[4948]: E0312 00:08:28.313025 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 00:08:28 crc kubenswrapper[4948]: I0312 00:08:28.383471 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:28 crc kubenswrapper[4948]: I0312 00:08:28.383546 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:28 crc kubenswrapper[4948]: I0312 00:08:28.383569 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:28 crc kubenswrapper[4948]: I0312 00:08:28.383598 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:28 crc kubenswrapper[4948]: I0312 00:08:28.383619 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:28Z","lastTransitionTime":"2026-03-12T00:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:28 crc kubenswrapper[4948]: I0312 00:08:28.486569 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:28 crc kubenswrapper[4948]: I0312 00:08:28.486628 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:28 crc kubenswrapper[4948]: I0312 00:08:28.486644 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:28 crc kubenswrapper[4948]: I0312 00:08:28.486667 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:28 crc kubenswrapper[4948]: I0312 00:08:28.486684 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:28Z","lastTransitionTime":"2026-03-12T00:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:28 crc kubenswrapper[4948]: I0312 00:08:28.589540 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:28 crc kubenswrapper[4948]: I0312 00:08:28.589599 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:28 crc kubenswrapper[4948]: I0312 00:08:28.589618 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:28 crc kubenswrapper[4948]: I0312 00:08:28.589646 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:28 crc kubenswrapper[4948]: I0312 00:08:28.589668 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:28Z","lastTransitionTime":"2026-03-12T00:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:28 crc kubenswrapper[4948]: I0312 00:08:28.692726 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:28 crc kubenswrapper[4948]: I0312 00:08:28.692809 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:28 crc kubenswrapper[4948]: I0312 00:08:28.692832 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:28 crc kubenswrapper[4948]: I0312 00:08:28.692863 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:28 crc kubenswrapper[4948]: I0312 00:08:28.692890 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:28Z","lastTransitionTime":"2026-03-12T00:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:28 crc kubenswrapper[4948]: I0312 00:08:28.795495 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:28 crc kubenswrapper[4948]: I0312 00:08:28.795586 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:28 crc kubenswrapper[4948]: I0312 00:08:28.795603 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:28 crc kubenswrapper[4948]: I0312 00:08:28.795626 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:28 crc kubenswrapper[4948]: I0312 00:08:28.795644 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:28Z","lastTransitionTime":"2026-03-12T00:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:28 crc kubenswrapper[4948]: I0312 00:08:28.898777 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:28 crc kubenswrapper[4948]: I0312 00:08:28.898827 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:28 crc kubenswrapper[4948]: I0312 00:08:28.898844 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:28 crc kubenswrapper[4948]: I0312 00:08:28.898867 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:28 crc kubenswrapper[4948]: I0312 00:08:28.898883 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:28Z","lastTransitionTime":"2026-03-12T00:08:28Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:29 crc kubenswrapper[4948]: I0312 00:08:29.001673 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:29 crc kubenswrapper[4948]: I0312 00:08:29.001716 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:29 crc kubenswrapper[4948]: I0312 00:08:29.001727 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:29 crc kubenswrapper[4948]: I0312 00:08:29.001744 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:29 crc kubenswrapper[4948]: I0312 00:08:29.001756 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:29Z","lastTransitionTime":"2026-03-12T00:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:29 crc kubenswrapper[4948]: I0312 00:08:29.105085 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:29 crc kubenswrapper[4948]: I0312 00:08:29.105153 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:29 crc kubenswrapper[4948]: I0312 00:08:29.105170 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:29 crc kubenswrapper[4948]: I0312 00:08:29.105191 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:29 crc kubenswrapper[4948]: I0312 00:08:29.105207 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:29Z","lastTransitionTime":"2026-03-12T00:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:29 crc kubenswrapper[4948]: I0312 00:08:29.208568 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:29 crc kubenswrapper[4948]: I0312 00:08:29.208872 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:29 crc kubenswrapper[4948]: I0312 00:08:29.208888 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:29 crc kubenswrapper[4948]: I0312 00:08:29.208912 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:29 crc kubenswrapper[4948]: I0312 00:08:29.208930 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:29Z","lastTransitionTime":"2026-03-12T00:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:29 crc kubenswrapper[4948]: I0312 00:08:29.311359 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:29 crc kubenswrapper[4948]: I0312 00:08:29.311406 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:29 crc kubenswrapper[4948]: I0312 00:08:29.311423 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:29 crc kubenswrapper[4948]: I0312 00:08:29.311446 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:29 crc kubenswrapper[4948]: I0312 00:08:29.311464 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:29Z","lastTransitionTime":"2026-03-12T00:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:29 crc kubenswrapper[4948]: I0312 00:08:29.415430 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:29 crc kubenswrapper[4948]: I0312 00:08:29.415494 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:29 crc kubenswrapper[4948]: I0312 00:08:29.415512 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:29 crc kubenswrapper[4948]: I0312 00:08:29.415536 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:29 crc kubenswrapper[4948]: I0312 00:08:29.415552 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:29Z","lastTransitionTime":"2026-03-12T00:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:29 crc kubenswrapper[4948]: I0312 00:08:29.518478 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:29 crc kubenswrapper[4948]: I0312 00:08:29.518539 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:29 crc kubenswrapper[4948]: I0312 00:08:29.518557 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:29 crc kubenswrapper[4948]: I0312 00:08:29.518582 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:29 crc kubenswrapper[4948]: I0312 00:08:29.518599 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:29Z","lastTransitionTime":"2026-03-12T00:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:29 crc kubenswrapper[4948]: I0312 00:08:29.621722 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:29 crc kubenswrapper[4948]: I0312 00:08:29.621795 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:29 crc kubenswrapper[4948]: I0312 00:08:29.621814 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:29 crc kubenswrapper[4948]: I0312 00:08:29.621839 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:29 crc kubenswrapper[4948]: I0312 00:08:29.621858 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:29Z","lastTransitionTime":"2026-03-12T00:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:29 crc kubenswrapper[4948]: I0312 00:08:29.724781 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:29 crc kubenswrapper[4948]: I0312 00:08:29.724838 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:29 crc kubenswrapper[4948]: I0312 00:08:29.724854 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:29 crc kubenswrapper[4948]: I0312 00:08:29.724877 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:29 crc kubenswrapper[4948]: I0312 00:08:29.724893 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:29Z","lastTransitionTime":"2026-03-12T00:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:29 crc kubenswrapper[4948]: I0312 00:08:29.827721 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:29 crc kubenswrapper[4948]: I0312 00:08:29.827770 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:29 crc kubenswrapper[4948]: I0312 00:08:29.827787 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:29 crc kubenswrapper[4948]: I0312 00:08:29.827808 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:29 crc kubenswrapper[4948]: I0312 00:08:29.827824 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:29Z","lastTransitionTime":"2026-03-12T00:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:29 crc kubenswrapper[4948]: I0312 00:08:29.931013 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:29 crc kubenswrapper[4948]: I0312 00:08:29.931076 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:29 crc kubenswrapper[4948]: I0312 00:08:29.931093 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:29 crc kubenswrapper[4948]: I0312 00:08:29.931118 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:29 crc kubenswrapper[4948]: I0312 00:08:29.931135 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:29Z","lastTransitionTime":"2026-03-12T00:08:29Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:30 crc kubenswrapper[4948]: I0312 00:08:30.034776 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:30 crc kubenswrapper[4948]: I0312 00:08:30.035212 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:30 crc kubenswrapper[4948]: I0312 00:08:30.035428 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:30 crc kubenswrapper[4948]: I0312 00:08:30.035601 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:30 crc kubenswrapper[4948]: I0312 00:08:30.035794 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:30Z","lastTransitionTime":"2026-03-12T00:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:30 crc kubenswrapper[4948]: I0312 00:08:30.042584 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 00:08:30 crc kubenswrapper[4948]: E0312 00:08:30.042879 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 00:08:46.042807469 +0000 UTC m=+125.498411237 (durationBeforeRetry 16s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:08:30 crc kubenswrapper[4948]: I0312 00:08:30.042999 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:08:30 crc kubenswrapper[4948]: I0312 00:08:30.043077 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:08:30 crc kubenswrapper[4948]: I0312 00:08:30.043183 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:08:30 crc kubenswrapper[4948]: E0312 00:08:30.043225 4948 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 12 00:08:30 crc kubenswrapper[4948]: E0312 00:08:30.043374 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-12 00:08:46.043344335 +0000 UTC m=+125.498948103 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 12 00:08:30 crc kubenswrapper[4948]: E0312 00:08:30.043413 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 12 00:08:30 crc kubenswrapper[4948]: E0312 00:08:30.043442 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 12 00:08:30 crc kubenswrapper[4948]: E0312 00:08:30.043459 4948 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 00:08:30 crc kubenswrapper[4948]: E0312 00:08:30.043481 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 12 00:08:30 crc kubenswrapper[4948]: E0312 00:08:30.043518 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 12 00:08:30 crc kubenswrapper[4948]: E0312 00:08:30.043526 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-12 00:08:46.043503663 +0000 UTC m=+125.499107441 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 00:08:30 crc kubenswrapper[4948]: I0312 00:08:30.043241 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:08:30 crc kubenswrapper[4948]: E0312 00:08:30.043544 4948 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 00:08:30 crc kubenswrapper[4948]: E0312 00:08:30.043629 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-12 00:08:46.043616438 +0000 UTC m=+125.499220206 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 00:08:30 crc kubenswrapper[4948]: E0312 00:08:30.043541 4948 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 12 00:08:30 crc kubenswrapper[4948]: E0312 00:08:30.043683 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-12 00:08:46.0436727 +0000 UTC m=+125.499276468 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 12 00:08:30 crc kubenswrapper[4948]: I0312 00:08:30.139523 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:30 crc kubenswrapper[4948]: I0312 00:08:30.139599 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:30 crc kubenswrapper[4948]: I0312 00:08:30.139624 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:30 crc kubenswrapper[4948]: I0312 00:08:30.139654 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:30 crc kubenswrapper[4948]: I0312 00:08:30.139678 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:30Z","lastTransitionTime":"2026-03-12T00:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:30 crc kubenswrapper[4948]: I0312 00:08:30.144725 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/be985ff1-de8a-431f-a230-67894ebe52d5-metrics-certs\") pod \"network-metrics-daemon-mqmtd\" (UID: \"be985ff1-de8a-431f-a230-67894ebe52d5\") " pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:08:30 crc kubenswrapper[4948]: E0312 00:08:30.144987 4948 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 12 00:08:30 crc kubenswrapper[4948]: E0312 00:08:30.145061 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/be985ff1-de8a-431f-a230-67894ebe52d5-metrics-certs podName:be985ff1-de8a-431f-a230-67894ebe52d5 nodeName:}" failed. No retries permitted until 2026-03-12 00:08:46.145036481 +0000 UTC m=+125.600640249 (durationBeforeRetry 16s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/be985ff1-de8a-431f-a230-67894ebe52d5-metrics-certs") pod "network-metrics-daemon-mqmtd" (UID: "be985ff1-de8a-431f-a230-67894ebe52d5") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 12 00:08:30 crc kubenswrapper[4948]: I0312 00:08:30.242211 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:30 crc kubenswrapper[4948]: I0312 00:08:30.242275 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:30 crc kubenswrapper[4948]: I0312 00:08:30.242297 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:30 crc kubenswrapper[4948]: I0312 00:08:30.242377 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:30 crc kubenswrapper[4948]: I0312 00:08:30.242400 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:30Z","lastTransitionTime":"2026-03-12T00:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:30 crc kubenswrapper[4948]: I0312 00:08:30.312065 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:08:30 crc kubenswrapper[4948]: E0312 00:08:30.312236 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 00:08:30 crc kubenswrapper[4948]: I0312 00:08:30.312496 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:08:30 crc kubenswrapper[4948]: I0312 00:08:30.312551 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:08:30 crc kubenswrapper[4948]: E0312 00:08:30.312624 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mqmtd" podUID="be985ff1-de8a-431f-a230-67894ebe52d5" Mar 12 00:08:30 crc kubenswrapper[4948]: E0312 00:08:30.312753 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 00:08:30 crc kubenswrapper[4948]: I0312 00:08:30.312777 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:08:30 crc kubenswrapper[4948]: E0312 00:08:30.312914 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 00:08:30 crc kubenswrapper[4948]: I0312 00:08:30.345675 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:30 crc kubenswrapper[4948]: I0312 00:08:30.345732 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:30 crc kubenswrapper[4948]: I0312 00:08:30.345748 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:30 crc kubenswrapper[4948]: I0312 00:08:30.345770 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:30 crc kubenswrapper[4948]: I0312 00:08:30.345787 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:30Z","lastTransitionTime":"2026-03-12T00:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:30 crc kubenswrapper[4948]: I0312 00:08:30.448639 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:30 crc kubenswrapper[4948]: I0312 00:08:30.448702 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:30 crc kubenswrapper[4948]: I0312 00:08:30.448729 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:30 crc kubenswrapper[4948]: I0312 00:08:30.448759 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:30 crc kubenswrapper[4948]: I0312 00:08:30.448782 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:30Z","lastTransitionTime":"2026-03-12T00:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:30 crc kubenswrapper[4948]: I0312 00:08:30.551232 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:30 crc kubenswrapper[4948]: I0312 00:08:30.551287 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:30 crc kubenswrapper[4948]: I0312 00:08:30.551331 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:30 crc kubenswrapper[4948]: I0312 00:08:30.551356 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:30 crc kubenswrapper[4948]: I0312 00:08:30.551374 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:30Z","lastTransitionTime":"2026-03-12T00:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:30 crc kubenswrapper[4948]: I0312 00:08:30.654337 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:30 crc kubenswrapper[4948]: I0312 00:08:30.654413 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:30 crc kubenswrapper[4948]: I0312 00:08:30.654440 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:30 crc kubenswrapper[4948]: I0312 00:08:30.654467 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:30 crc kubenswrapper[4948]: I0312 00:08:30.654514 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:30Z","lastTransitionTime":"2026-03-12T00:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:30 crc kubenswrapper[4948]: I0312 00:08:30.758622 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:30 crc kubenswrapper[4948]: I0312 00:08:30.758681 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:30 crc kubenswrapper[4948]: I0312 00:08:30.758705 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:30 crc kubenswrapper[4948]: I0312 00:08:30.758735 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:30 crc kubenswrapper[4948]: I0312 00:08:30.758760 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:30Z","lastTransitionTime":"2026-03-12T00:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:30 crc kubenswrapper[4948]: I0312 00:08:30.861896 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:30 crc kubenswrapper[4948]: I0312 00:08:30.861960 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:30 crc kubenswrapper[4948]: I0312 00:08:30.861977 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:30 crc kubenswrapper[4948]: I0312 00:08:30.862003 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:30 crc kubenswrapper[4948]: I0312 00:08:30.862020 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:30Z","lastTransitionTime":"2026-03-12T00:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:30 crc kubenswrapper[4948]: I0312 00:08:30.965232 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:30 crc kubenswrapper[4948]: I0312 00:08:30.965334 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:30 crc kubenswrapper[4948]: I0312 00:08:30.965353 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:30 crc kubenswrapper[4948]: I0312 00:08:30.965382 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:30 crc kubenswrapper[4948]: I0312 00:08:30.965399 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:30Z","lastTransitionTime":"2026-03-12T00:08:30Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.068139 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.068178 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.068187 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.068204 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.068214 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:31Z","lastTransitionTime":"2026-03-12T00:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.170369 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.170431 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.170449 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.170472 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.170491 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:31Z","lastTransitionTime":"2026-03-12T00:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.274034 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.274101 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.274118 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.274144 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.274193 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:31Z","lastTransitionTime":"2026-03-12T00:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.335392 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mqmtd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be985ff1-de8a-431f-a230-67894ebe52d5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmsz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmsz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mqmtd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:31Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.367282 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15117613-9fad-48c7-98c4-a2d84502ded9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4036d7131db00afd8eab8e093626b688297dc776b5af80c3c38b51da7124085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84af4cdd213d90636a889a8a8a3aaca3dee1e2677368efa0ffc768940c2cc994\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e481ace9e84b6aa66cbdda3ee8425c4120a665915d2c70ae4faf60bba6205e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9eadefd0babc119f776d48c9d2c9e4bced58db685af9a8e97f25e6a9a2e3c89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9d5e77753fda70350353daa0967257dd4929dba4440a95495ab4c2a0659aef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cad943edb9a8d445b33a6e70d85b475c1ccba460a6056bccae94fe60fd6b73f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://588e8a26bafbb970547438b573ab34f111ef18d8134df2b3fff513c9cca9a516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://588e8a26bafbb970547438b573ab34f111ef18d8134df2b3fff513c9cca9a516\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T00:08:25Z\\\",\\\"message\\\":\\\"umn _uuid == {f6d604c1-9711-4e25-be6c-79ec28bbad1b}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0312 00:08:25.737942 6942 obj_retry.go:551] Creating *factory.egressNode crc took: 2.187304ms\\\\nI0312 00:08:25.737970 6942 factory.go:1336] Added *v1.Node event handler 7\\\\nI0312 00:08:25.738003 6942 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI0312 00:08:25.738210 6942 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0312 00:08:25.738274 6942 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0312 00:08:25.738325 6942 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0312 00:08:25.738333 6942 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0312 00:08:25.738357 6942 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0312 00:08:25.738373 6942 factory.go:656] Stopping watch factory\\\\nI0312 00:08:25.738386 6942 ovnkube.go:599] Stopped ovnkube\\\\nI0312 00:08:25.738410 6942 handler.go:208] Removed *v1.Node event handler 2\\\\nI0312 00:08:25.738414 6942 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0312 00:08:25.738430 6942 handler.go:208] Removed *v1.Node event handler 7\\\\nI0312 00:08:25.738442 6942 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nF0312 00:08:25.738550 6942 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:24Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-f7mp5_openshift-ovn-kubernetes(15117613-9fad-48c7-98c4-a2d84502ded9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46fa6140f4f4bd32e298fa36f6d7af6890029f9bc3080d2b4d94be8307346de8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7mp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:31Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.377568 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.377624 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.377643 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.377668 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.377685 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:31Z","lastTransitionTime":"2026-03-12T00:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.387578 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:31Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.407175 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:31Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.426876 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8d58ee45d75284e030700fa526b2c39197a26f6a391a7417f25a026f356b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:31Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.452423 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:31Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.477131 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfp2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"662dcbe7-7cf8-42d4-9a5e-345c32a49972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff3f3c69dcb28c7b0fd85f9e0c33b3339a1f6159dbf10bcf0efbdce4f7728c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a57d476b5b00110168878bd1c19c32ef122525c0259ac0a70de31c8220996d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4a57d476b5b00110168878bd1c19c32ef122525c0259ac0a70de31c8220996d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bb429b94a1286d8aac77cb77112c7163f7aa50a1c9cab997f8d088c100887c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0bb429b94a1286d8aac77cb77112c7163f7aa50a1c9cab997f8d088c100887c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a77485fd9e246f190724b86663e0dbe81146e1fcc51f468dd05f23c1db878fd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a77485fd9e246f190724b86663e0dbe81146e1fcc51f468dd05f23c1db878fd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adaa20d9a4898057802a8c9759ab3cd485963180bfc4f6b926bf6d2c886bcf27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adaa20d9a4898057802a8c9759ab3cd485963180bfc4f6b926bf6d2c886bcf27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://998fbb934f01c9ecd315f78e776199d4fd538106f8373f09b730d2df23ac9e3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://998fbb934f01c9ecd315f78e776199d4fd538106f8373f09b730d2df23ac9e3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46ebe18804ff07eb0e922a6cc1c39ff93875302d231222642b211e3527fe6373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46ebe18804ff07eb0e922a6cc1c39ff93875302d231222642b211e3527fe6373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfp2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:31Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.480464 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.480515 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.480533 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.480557 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.480575 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:31Z","lastTransitionTime":"2026-03-12T00:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.497823 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e483c2b-08f0-4e92-8e4a-b7281f30af3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d6839c99ab958ad1f40636c70e29bc285292fdbee1e910cdf7f8ab43ca7e8db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd9jv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82de8f2b633f1b6c295149d077bfd2b308f5acaa429acf411ed16f456536d0ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd9jv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4xwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:31Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.516870 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5139686878eb07d3abced92d0ccfe05be8907580ed75233b72fc2ad3415a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15cbe3fa75b69a7176803295e16a314a8bc73f5797c874124d54e76d47fa1064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:31Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.537690 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwpqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b7580c2-5c23-4c67-807a-ea97a3df9398\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d07c603619f858177d4b007e785208853abcfe2dc6c1af435309030fa409010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8lxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwpqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:31Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.553403 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-njc9c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0d6440c-4524-42ea-b18a-0ddc4662f672\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f4d0d9599ad62aaba35ba1414a806f3a163e29a3c34fd893ef5b14eae003591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7wn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-njc9c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:31Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.575189 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2efff53-8e12-41a1-bcb1-9a1d08c42a4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1cb328c8431b39dac735d49726a2eae40a367849b2c146bb2e49d02958f0e5a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ad13662849131222d5c8ba1041b81fa75bcb1fcbfa2e71c0c751f04b532a5b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8cd4faf47b25395ffd42d4bf5416ea29dcaa3ec380d6d6d6a9f13321d35195a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://73834c8db23f6536034dce09e38b56665211238280bf9a4e1b13e74d81245a77\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73834c8db23f6536034dce09e38b56665211238280bf9a4e1b13e74d81245a77\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T00:07:50Z\\\",\\\"message\\\":\\\"le observer\\\\nW0312 00:07:49.927227 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0312 00:07:49.927347 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 00:07:49.928000 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2511358960/tls.crt::/tmp/serving-cert-2511358960/tls.key\\\\\\\"\\\\nI0312 00:07:50.321455 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0312 00:07:50.324479 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0312 00:07:50.324511 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0312 00:07:50.324546 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0312 00:07:50.324556 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0312 00:07:50.331591 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0312 00:07:50.331609 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0312 00:07:50.331634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 00:07:50.331642 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 00:07:50.331651 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0312 00:07:50.331658 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0312 00:07:50.331664 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0312 00:07:50.331670 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0312 00:07:50.332917 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:07:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=kube-apiserver-check-endpoints pod=kube-apiserver-crc_openshift-kube-apiserver(f4b27818a5e8e43d0dc095d08835c792)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d56e32f9cb96ad103bab80d6a730b70d3387264a25a7a61ef1f0726e7290d6c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e42f2e94463ac8a3d7980a25122062da9ee87fcdc1f10e260ff9974b2e42359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e42f2e94463ac8a3d7980a25122062da9ee87fcdc1f10e260ff9974b2e42359\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:06:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:31Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.583380 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.583506 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.583580 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.583611 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.583668 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:31Z","lastTransitionTime":"2026-03-12T00:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.591556 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5a9ae9441886817563a5fc793bff7c9b8d77d406ceb1d1fb1eebd4307a1c64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:31Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.605988 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9s5fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f9c7f1-6593-4d92-854c-e6c9964a0435\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b8833cbaf3809463e18bfebc9d5bada5fb26823831cc742a3a27177662f60e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xf4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9s5fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:31Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.622640 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xdd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b5233d6-d2b8-42dc-b881-18e439ecc442\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0ea55d993c72b6d50a4829ecfff4adae7d0f625a90e40545c9e63201cdda29c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjpwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb23ff4bc484524e306561c0c574e2b2eee39fc80efcd9f44f1924fe42df1a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjpwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4xdd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:31Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.686395 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.686457 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.686474 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.686499 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.686517 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:31Z","lastTransitionTime":"2026-03-12T00:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.789240 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.789342 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.789365 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.789397 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.789422 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:31Z","lastTransitionTime":"2026-03-12T00:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.871380 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.871439 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.871461 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.871493 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.871519 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:31Z","lastTransitionTime":"2026-03-12T00:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:31 crc kubenswrapper[4948]: E0312 00:08:31.895587 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17825588-29cf-4059-bab7-1b042b9bf1f3\\\",\\\"systemUUID\\\":\\\"b18e4bf2-d37e-4b80-8da3-cd3803754ba2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:31Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.900796 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.900846 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.900863 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.900886 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.900902 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:31Z","lastTransitionTime":"2026-03-12T00:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:31 crc kubenswrapper[4948]: E0312 00:08:31.921346 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17825588-29cf-4059-bab7-1b042b9bf1f3\\\",\\\"systemUUID\\\":\\\"b18e4bf2-d37e-4b80-8da3-cd3803754ba2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:31Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.926289 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.926366 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.926385 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.926407 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.926423 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:31Z","lastTransitionTime":"2026-03-12T00:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:31 crc kubenswrapper[4948]: E0312 00:08:31.945161 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17825588-29cf-4059-bab7-1b042b9bf1f3\\\",\\\"systemUUID\\\":\\\"b18e4bf2-d37e-4b80-8da3-cd3803754ba2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:31Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.949884 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.950078 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.950218 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.950428 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.950587 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:31Z","lastTransitionTime":"2026-03-12T00:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:31 crc kubenswrapper[4948]: E0312 00:08:31.977105 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17825588-29cf-4059-bab7-1b042b9bf1f3\\\",\\\"systemUUID\\\":\\\"b18e4bf2-d37e-4b80-8da3-cd3803754ba2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:31Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.981733 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.981825 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.981850 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.981880 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:31 crc kubenswrapper[4948]: I0312 00:08:31.981900 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:31Z","lastTransitionTime":"2026-03-12T00:08:31Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:32 crc kubenswrapper[4948]: E0312 00:08:32.002170 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:31Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:31Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:31Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:31Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17825588-29cf-4059-bab7-1b042b9bf1f3\\\",\\\"systemUUID\\\":\\\"b18e4bf2-d37e-4b80-8da3-cd3803754ba2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:31Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:32 crc kubenswrapper[4948]: E0312 00:08:32.002412 4948 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 12 00:08:32 crc kubenswrapper[4948]: I0312 00:08:32.004513 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:32 crc kubenswrapper[4948]: I0312 00:08:32.004693 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:32 crc kubenswrapper[4948]: I0312 00:08:32.004719 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:32 crc kubenswrapper[4948]: I0312 00:08:32.004748 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:32 crc kubenswrapper[4948]: I0312 00:08:32.004773 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:32Z","lastTransitionTime":"2026-03-12T00:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:32 crc kubenswrapper[4948]: I0312 00:08:32.108084 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:32 crc kubenswrapper[4948]: I0312 00:08:32.108415 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:32 crc kubenswrapper[4948]: I0312 00:08:32.108555 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:32 crc kubenswrapper[4948]: I0312 00:08:32.108692 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:32 crc kubenswrapper[4948]: I0312 00:08:32.108809 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:32Z","lastTransitionTime":"2026-03-12T00:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:32 crc kubenswrapper[4948]: I0312 00:08:32.211989 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:32 crc kubenswrapper[4948]: I0312 00:08:32.212284 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:32 crc kubenswrapper[4948]: I0312 00:08:32.212481 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:32 crc kubenswrapper[4948]: I0312 00:08:32.212621 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:32 crc kubenswrapper[4948]: I0312 00:08:32.212749 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:32Z","lastTransitionTime":"2026-03-12T00:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:32 crc kubenswrapper[4948]: I0312 00:08:32.312034 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:08:32 crc kubenswrapper[4948]: E0312 00:08:32.312543 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mqmtd" podUID="be985ff1-de8a-431f-a230-67894ebe52d5" Mar 12 00:08:32 crc kubenswrapper[4948]: I0312 00:08:32.312097 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:08:32 crc kubenswrapper[4948]: E0312 00:08:32.312937 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 00:08:32 crc kubenswrapper[4948]: I0312 00:08:32.312035 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:08:32 crc kubenswrapper[4948]: E0312 00:08:32.313263 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 00:08:32 crc kubenswrapper[4948]: I0312 00:08:32.312094 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:08:32 crc kubenswrapper[4948]: E0312 00:08:32.313819 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 00:08:32 crc kubenswrapper[4948]: I0312 00:08:32.316345 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:32 crc kubenswrapper[4948]: I0312 00:08:32.316414 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:32 crc kubenswrapper[4948]: I0312 00:08:32.316433 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:32 crc kubenswrapper[4948]: I0312 00:08:32.316457 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:32 crc kubenswrapper[4948]: I0312 00:08:32.316474 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:32Z","lastTransitionTime":"2026-03-12T00:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:32 crc kubenswrapper[4948]: I0312 00:08:32.419704 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:32 crc kubenswrapper[4948]: I0312 00:08:32.419763 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:32 crc kubenswrapper[4948]: I0312 00:08:32.419781 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:32 crc kubenswrapper[4948]: I0312 00:08:32.419804 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:32 crc kubenswrapper[4948]: I0312 00:08:32.419822 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:32Z","lastTransitionTime":"2026-03-12T00:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:32 crc kubenswrapper[4948]: I0312 00:08:32.523145 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:32 crc kubenswrapper[4948]: I0312 00:08:32.523197 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:32 crc kubenswrapper[4948]: I0312 00:08:32.523214 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:32 crc kubenswrapper[4948]: I0312 00:08:32.523237 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:32 crc kubenswrapper[4948]: I0312 00:08:32.523253 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:32Z","lastTransitionTime":"2026-03-12T00:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:32 crc kubenswrapper[4948]: I0312 00:08:32.626461 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:32 crc kubenswrapper[4948]: I0312 00:08:32.626812 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:32 crc kubenswrapper[4948]: I0312 00:08:32.626956 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:32 crc kubenswrapper[4948]: I0312 00:08:32.627095 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:32 crc kubenswrapper[4948]: I0312 00:08:32.627213 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:32Z","lastTransitionTime":"2026-03-12T00:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:32 crc kubenswrapper[4948]: I0312 00:08:32.730435 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:32 crc kubenswrapper[4948]: I0312 00:08:32.730508 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:32 crc kubenswrapper[4948]: I0312 00:08:32.730526 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:32 crc kubenswrapper[4948]: I0312 00:08:32.730550 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:32 crc kubenswrapper[4948]: I0312 00:08:32.730570 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:32Z","lastTransitionTime":"2026-03-12T00:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:32 crc kubenswrapper[4948]: I0312 00:08:32.833977 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:32 crc kubenswrapper[4948]: I0312 00:08:32.834050 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:32 crc kubenswrapper[4948]: I0312 00:08:32.834069 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:32 crc kubenswrapper[4948]: I0312 00:08:32.834093 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:32 crc kubenswrapper[4948]: I0312 00:08:32.834110 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:32Z","lastTransitionTime":"2026-03-12T00:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:32 crc kubenswrapper[4948]: I0312 00:08:32.937124 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:32 crc kubenswrapper[4948]: I0312 00:08:32.937941 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:32 crc kubenswrapper[4948]: I0312 00:08:32.938080 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:32 crc kubenswrapper[4948]: I0312 00:08:32.938203 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:32 crc kubenswrapper[4948]: I0312 00:08:32.938374 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:32Z","lastTransitionTime":"2026-03-12T00:08:32Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:33 crc kubenswrapper[4948]: I0312 00:08:33.042136 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:33 crc kubenswrapper[4948]: I0312 00:08:33.042442 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:33 crc kubenswrapper[4948]: I0312 00:08:33.042578 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:33 crc kubenswrapper[4948]: I0312 00:08:33.042711 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:33 crc kubenswrapper[4948]: I0312 00:08:33.042824 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:33Z","lastTransitionTime":"2026-03-12T00:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:33 crc kubenswrapper[4948]: I0312 00:08:33.145209 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:33 crc kubenswrapper[4948]: I0312 00:08:33.145274 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:33 crc kubenswrapper[4948]: I0312 00:08:33.145293 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:33 crc kubenswrapper[4948]: I0312 00:08:33.145349 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:33 crc kubenswrapper[4948]: I0312 00:08:33.145368 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:33Z","lastTransitionTime":"2026-03-12T00:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:33 crc kubenswrapper[4948]: I0312 00:08:33.247280 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:33 crc kubenswrapper[4948]: I0312 00:08:33.247871 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:33 crc kubenswrapper[4948]: I0312 00:08:33.248029 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:33 crc kubenswrapper[4948]: I0312 00:08:33.248173 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:33 crc kubenswrapper[4948]: I0312 00:08:33.248296 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:33Z","lastTransitionTime":"2026-03-12T00:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:33 crc kubenswrapper[4948]: I0312 00:08:33.351863 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:33 crc kubenswrapper[4948]: I0312 00:08:33.351935 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:33 crc kubenswrapper[4948]: I0312 00:08:33.351954 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:33 crc kubenswrapper[4948]: I0312 00:08:33.351979 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:33 crc kubenswrapper[4948]: I0312 00:08:33.351996 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:33Z","lastTransitionTime":"2026-03-12T00:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:33 crc kubenswrapper[4948]: I0312 00:08:33.455041 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:33 crc kubenswrapper[4948]: I0312 00:08:33.455093 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:33 crc kubenswrapper[4948]: I0312 00:08:33.455112 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:33 crc kubenswrapper[4948]: I0312 00:08:33.455138 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:33 crc kubenswrapper[4948]: I0312 00:08:33.455154 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:33Z","lastTransitionTime":"2026-03-12T00:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:33 crc kubenswrapper[4948]: I0312 00:08:33.558862 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:33 crc kubenswrapper[4948]: I0312 00:08:33.558904 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:33 crc kubenswrapper[4948]: I0312 00:08:33.558920 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:33 crc kubenswrapper[4948]: I0312 00:08:33.558941 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:33 crc kubenswrapper[4948]: I0312 00:08:33.558956 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:33Z","lastTransitionTime":"2026-03-12T00:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:33 crc kubenswrapper[4948]: I0312 00:08:33.662415 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:33 crc kubenswrapper[4948]: I0312 00:08:33.662486 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:33 crc kubenswrapper[4948]: I0312 00:08:33.662509 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:33 crc kubenswrapper[4948]: I0312 00:08:33.662542 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:33 crc kubenswrapper[4948]: I0312 00:08:33.662565 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:33Z","lastTransitionTime":"2026-03-12T00:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:33 crc kubenswrapper[4948]: I0312 00:08:33.764909 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:33 crc kubenswrapper[4948]: I0312 00:08:33.764971 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:33 crc kubenswrapper[4948]: I0312 00:08:33.764986 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:33 crc kubenswrapper[4948]: I0312 00:08:33.765008 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:33 crc kubenswrapper[4948]: I0312 00:08:33.765022 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:33Z","lastTransitionTime":"2026-03-12T00:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:33 crc kubenswrapper[4948]: I0312 00:08:33.867365 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:33 crc kubenswrapper[4948]: I0312 00:08:33.867425 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:33 crc kubenswrapper[4948]: I0312 00:08:33.867443 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:33 crc kubenswrapper[4948]: I0312 00:08:33.867467 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:33 crc kubenswrapper[4948]: I0312 00:08:33.867486 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:33Z","lastTransitionTime":"2026-03-12T00:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:33 crc kubenswrapper[4948]: I0312 00:08:33.970609 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:33 crc kubenswrapper[4948]: I0312 00:08:33.970683 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:33 crc kubenswrapper[4948]: I0312 00:08:33.970699 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:33 crc kubenswrapper[4948]: I0312 00:08:33.970723 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:33 crc kubenswrapper[4948]: I0312 00:08:33.970740 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:33Z","lastTransitionTime":"2026-03-12T00:08:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:34 crc kubenswrapper[4948]: I0312 00:08:34.073096 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:34 crc kubenswrapper[4948]: I0312 00:08:34.073157 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:34 crc kubenswrapper[4948]: I0312 00:08:34.073174 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:34 crc kubenswrapper[4948]: I0312 00:08:34.073198 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:34 crc kubenswrapper[4948]: I0312 00:08:34.073216 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:34Z","lastTransitionTime":"2026-03-12T00:08:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:34 crc kubenswrapper[4948]: I0312 00:08:34.177202 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:34 crc kubenswrapper[4948]: I0312 00:08:34.177265 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:34 crc kubenswrapper[4948]: I0312 00:08:34.177284 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:34 crc kubenswrapper[4948]: I0312 00:08:34.177345 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:34 crc kubenswrapper[4948]: I0312 00:08:34.177366 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:34Z","lastTransitionTime":"2026-03-12T00:08:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:34 crc kubenswrapper[4948]: I0312 00:08:34.280990 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:34 crc kubenswrapper[4948]: I0312 00:08:34.281048 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:34 crc kubenswrapper[4948]: I0312 00:08:34.281059 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:34 crc kubenswrapper[4948]: I0312 00:08:34.281081 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:34 crc kubenswrapper[4948]: I0312 00:08:34.281095 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:34Z","lastTransitionTime":"2026-03-12T00:08:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:34 crc kubenswrapper[4948]: I0312 00:08:34.311866 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:08:34 crc kubenswrapper[4948]: I0312 00:08:34.311922 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:08:34 crc kubenswrapper[4948]: I0312 00:08:34.312059 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:08:34 crc kubenswrapper[4948]: E0312 00:08:34.312060 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 00:08:34 crc kubenswrapper[4948]: E0312 00:08:34.312264 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 00:08:34 crc kubenswrapper[4948]: I0312 00:08:34.312398 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:08:34 crc kubenswrapper[4948]: E0312 00:08:34.312872 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mqmtd" podUID="be985ff1-de8a-431f-a230-67894ebe52d5" Mar 12 00:08:34 crc kubenswrapper[4948]: E0312 00:08:34.313059 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 00:08:34 crc kubenswrapper[4948]: I0312 00:08:34.330868 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler/openshift-kube-scheduler-crc"] Mar 12 00:08:34 crc kubenswrapper[4948]: I0312 00:08:34.385040 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:34 crc kubenswrapper[4948]: I0312 00:08:34.385087 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:34 crc kubenswrapper[4948]: I0312 00:08:34.385106 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:34 crc kubenswrapper[4948]: I0312 00:08:34.385131 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:34 crc kubenswrapper[4948]: I0312 00:08:34.385151 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:34Z","lastTransitionTime":"2026-03-12T00:08:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:34 crc kubenswrapper[4948]: I0312 00:08:34.488046 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:34 crc kubenswrapper[4948]: I0312 00:08:34.488114 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:34 crc kubenswrapper[4948]: I0312 00:08:34.488138 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:34 crc kubenswrapper[4948]: I0312 00:08:34.488169 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:34 crc kubenswrapper[4948]: I0312 00:08:34.488194 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:34Z","lastTransitionTime":"2026-03-12T00:08:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:34 crc kubenswrapper[4948]: I0312 00:08:34.591586 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:34 crc kubenswrapper[4948]: I0312 00:08:34.591684 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:34 crc kubenswrapper[4948]: I0312 00:08:34.591702 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:34 crc kubenswrapper[4948]: I0312 00:08:34.591758 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:34 crc kubenswrapper[4948]: I0312 00:08:34.591776 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:34Z","lastTransitionTime":"2026-03-12T00:08:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:34 crc kubenswrapper[4948]: I0312 00:08:34.694582 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:34 crc kubenswrapper[4948]: I0312 00:08:34.694689 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:34 crc kubenswrapper[4948]: I0312 00:08:34.694708 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:34 crc kubenswrapper[4948]: I0312 00:08:34.694781 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:34 crc kubenswrapper[4948]: I0312 00:08:34.694805 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:34Z","lastTransitionTime":"2026-03-12T00:08:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:34 crc kubenswrapper[4948]: I0312 00:08:34.797745 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:34 crc kubenswrapper[4948]: I0312 00:08:34.797812 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:34 crc kubenswrapper[4948]: I0312 00:08:34.797829 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:34 crc kubenswrapper[4948]: I0312 00:08:34.797854 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:34 crc kubenswrapper[4948]: I0312 00:08:34.797870 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:34Z","lastTransitionTime":"2026-03-12T00:08:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:34 crc kubenswrapper[4948]: I0312 00:08:34.902194 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:34 crc kubenswrapper[4948]: I0312 00:08:34.902283 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:34 crc kubenswrapper[4948]: I0312 00:08:34.902351 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:34 crc kubenswrapper[4948]: I0312 00:08:34.902380 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:34 crc kubenswrapper[4948]: I0312 00:08:34.902398 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:34Z","lastTransitionTime":"2026-03-12T00:08:34Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:35 crc kubenswrapper[4948]: I0312 00:08:35.006168 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:35 crc kubenswrapper[4948]: I0312 00:08:35.006230 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:35 crc kubenswrapper[4948]: I0312 00:08:35.006249 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:35 crc kubenswrapper[4948]: I0312 00:08:35.006274 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:35 crc kubenswrapper[4948]: I0312 00:08:35.006293 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:35Z","lastTransitionTime":"2026-03-12T00:08:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:35 crc kubenswrapper[4948]: I0312 00:08:35.109329 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:35 crc kubenswrapper[4948]: I0312 00:08:35.109393 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:35 crc kubenswrapper[4948]: I0312 00:08:35.109410 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:35 crc kubenswrapper[4948]: I0312 00:08:35.109434 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:35 crc kubenswrapper[4948]: I0312 00:08:35.109453 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:35Z","lastTransitionTime":"2026-03-12T00:08:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:35 crc kubenswrapper[4948]: I0312 00:08:35.213055 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:35 crc kubenswrapper[4948]: I0312 00:08:35.213123 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:35 crc kubenswrapper[4948]: I0312 00:08:35.213145 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:35 crc kubenswrapper[4948]: I0312 00:08:35.213170 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:35 crc kubenswrapper[4948]: I0312 00:08:35.213187 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:35Z","lastTransitionTime":"2026-03-12T00:08:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:35 crc kubenswrapper[4948]: I0312 00:08:35.316644 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:35 crc kubenswrapper[4948]: I0312 00:08:35.316719 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:35 crc kubenswrapper[4948]: I0312 00:08:35.316746 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:35 crc kubenswrapper[4948]: I0312 00:08:35.316785 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:35 crc kubenswrapper[4948]: I0312 00:08:35.316882 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:35Z","lastTransitionTime":"2026-03-12T00:08:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:35 crc kubenswrapper[4948]: I0312 00:08:35.419847 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:35 crc kubenswrapper[4948]: I0312 00:08:35.420022 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:35 crc kubenswrapper[4948]: I0312 00:08:35.420044 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:35 crc kubenswrapper[4948]: I0312 00:08:35.420069 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:35 crc kubenswrapper[4948]: I0312 00:08:35.420091 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:35Z","lastTransitionTime":"2026-03-12T00:08:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:35 crc kubenswrapper[4948]: I0312 00:08:35.523251 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:35 crc kubenswrapper[4948]: I0312 00:08:35.523356 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:35 crc kubenswrapper[4948]: I0312 00:08:35.523380 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:35 crc kubenswrapper[4948]: I0312 00:08:35.523406 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:35 crc kubenswrapper[4948]: I0312 00:08:35.523426 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:35Z","lastTransitionTime":"2026-03-12T00:08:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:35 crc kubenswrapper[4948]: I0312 00:08:35.627360 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:35 crc kubenswrapper[4948]: I0312 00:08:35.627442 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:35 crc kubenswrapper[4948]: I0312 00:08:35.627461 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:35 crc kubenswrapper[4948]: I0312 00:08:35.627492 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:35 crc kubenswrapper[4948]: I0312 00:08:35.627511 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:35Z","lastTransitionTime":"2026-03-12T00:08:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:35 crc kubenswrapper[4948]: I0312 00:08:35.730648 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:35 crc kubenswrapper[4948]: I0312 00:08:35.730722 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:35 crc kubenswrapper[4948]: I0312 00:08:35.730735 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:35 crc kubenswrapper[4948]: I0312 00:08:35.730764 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:35 crc kubenswrapper[4948]: I0312 00:08:35.730782 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:35Z","lastTransitionTime":"2026-03-12T00:08:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:35 crc kubenswrapper[4948]: I0312 00:08:35.833988 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:35 crc kubenswrapper[4948]: I0312 00:08:35.834036 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:35 crc kubenswrapper[4948]: I0312 00:08:35.834053 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:35 crc kubenswrapper[4948]: I0312 00:08:35.834081 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:35 crc kubenswrapper[4948]: I0312 00:08:35.834098 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:35Z","lastTransitionTime":"2026-03-12T00:08:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:35 crc kubenswrapper[4948]: I0312 00:08:35.936748 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:35 crc kubenswrapper[4948]: I0312 00:08:35.936808 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:35 crc kubenswrapper[4948]: I0312 00:08:35.936825 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:35 crc kubenswrapper[4948]: I0312 00:08:35.936849 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:35 crc kubenswrapper[4948]: I0312 00:08:35.936870 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:35Z","lastTransitionTime":"2026-03-12T00:08:35Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.040673 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.040738 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.040755 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.040779 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.040796 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:36Z","lastTransitionTime":"2026-03-12T00:08:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.144022 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.144102 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.144123 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.144153 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.144173 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:36Z","lastTransitionTime":"2026-03-12T00:08:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.247548 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.247643 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.247662 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.247688 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.247705 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:36Z","lastTransitionTime":"2026-03-12T00:08:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.311904 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.312013 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.312021 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.312215 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:08:36 crc kubenswrapper[4948]: E0312 00:08:36.312459 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 00:08:36 crc kubenswrapper[4948]: E0312 00:08:36.312626 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mqmtd" podUID="be985ff1-de8a-431f-a230-67894ebe52d5" Mar 12 00:08:36 crc kubenswrapper[4948]: E0312 00:08:36.312866 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 00:08:36 crc kubenswrapper[4948]: E0312 00:08:36.313018 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.314020 4948 scope.go:117] "RemoveContainer" containerID="73834c8db23f6536034dce09e38b56665211238280bf9a4e1b13e74d81245a77" Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.326882 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/kube-controller-manager-crc"] Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.349949 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.350004 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.350021 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.350044 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.350060 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:36Z","lastTransitionTime":"2026-03-12T00:08:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.452810 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.452864 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.452878 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.452902 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.452915 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:36Z","lastTransitionTime":"2026-03-12T00:08:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.555612 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.555672 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.555691 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.555718 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.555736 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:36Z","lastTransitionTime":"2026-03-12T00:08:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.658516 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.658580 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.658599 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.658622 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.658641 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:36Z","lastTransitionTime":"2026-03-12T00:08:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.723327 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.726123 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"f4b27818a5e8e43d0dc095d08835c792","Type":"ContainerStarted","Data":"8028fd7b3e6d26bf30666425c843933b8fb87c3b2a258d510608ec718eb4fe82"} Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.727114 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.746819 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:36Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.761188 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.761666 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.761684 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.761709 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.761726 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:36Z","lastTransitionTime":"2026-03-12T00:08:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.766989 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:36Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.786870 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8d58ee45d75284e030700fa526b2c39197a26f6a391a7417f25a026f356b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:36Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.803729 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mqmtd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be985ff1-de8a-431f-a230-67894ebe52d5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmsz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmsz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mqmtd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:36Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.836956 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15117613-9fad-48c7-98c4-a2d84502ded9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4036d7131db00afd8eab8e093626b688297dc776b5af80c3c38b51da7124085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84af4cdd213d90636a889a8a8a3aaca3dee1e2677368efa0ffc768940c2cc994\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e481ace9e84b6aa66cbdda3ee8425c4120a665915d2c70ae4faf60bba6205e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9eadefd0babc119f776d48c9d2c9e4bced58db685af9a8e97f25e6a9a2e3c89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9d5e77753fda70350353daa0967257dd4929dba4440a95495ab4c2a0659aef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cad943edb9a8d445b33a6e70d85b475c1ccba460a6056bccae94fe60fd6b73f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://588e8a26bafbb970547438b573ab34f111ef18d8134df2b3fff513c9cca9a516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://588e8a26bafbb970547438b573ab34f111ef18d8134df2b3fff513c9cca9a516\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T00:08:25Z\\\",\\\"message\\\":\\\"umn _uuid == {f6d604c1-9711-4e25-be6c-79ec28bbad1b}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0312 00:08:25.737942 6942 obj_retry.go:551] Creating *factory.egressNode crc took: 2.187304ms\\\\nI0312 00:08:25.737970 6942 factory.go:1336] Added *v1.Node event handler 7\\\\nI0312 00:08:25.738003 6942 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI0312 00:08:25.738210 6942 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0312 00:08:25.738274 6942 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0312 00:08:25.738325 6942 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0312 00:08:25.738333 6942 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0312 00:08:25.738357 6942 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0312 00:08:25.738373 6942 factory.go:656] Stopping watch factory\\\\nI0312 00:08:25.738386 6942 ovnkube.go:599] Stopped ovnkube\\\\nI0312 00:08:25.738410 6942 handler.go:208] Removed *v1.Node event handler 2\\\\nI0312 00:08:25.738414 6942 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0312 00:08:25.738430 6942 handler.go:208] Removed *v1.Node event handler 7\\\\nI0312 00:08:25.738442 6942 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nF0312 00:08:25.738550 6942 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:24Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-f7mp5_openshift-ovn-kubernetes(15117613-9fad-48c7-98c4-a2d84502ded9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46fa6140f4f4bd32e298fa36f6d7af6890029f9bc3080d2b4d94be8307346de8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7mp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:36Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.857387 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c835bb7a-4e06-4047-9fe6-e0c605b36bc6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc9760bd6ed0d092081fdd2ad33dda2d67f1305531640a35b9dce84b5d86a813\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f580535b16c31fb0ebadbdd9b7851ba2dce2f7d52bcad8c9cacc197f5e7dfad4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T00:07:08Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0312 00:06:43.492131 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0312 00:06:43.494442 1 observer_polling.go:159] Starting file observer\\\\nI0312 00:06:43.528056 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0312 00:06:43.536640 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0312 00:07:08.239474 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0312 00:07:08.239540 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb45482a404d6696cfbc3919997e53146b7b9dd1536056a55374574ae01ff193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://663bca87e44109c1d2bb59964e567d8b78fe2e9de39bcec0557cb43e39143add\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bf5b5d40a0cc76725ae2cf453089b31e75c304cf6037ef06b8ca8f7cff35a5a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:06:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:36Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.865039 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.865092 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.865116 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.865145 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.865166 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:36Z","lastTransitionTime":"2026-03-12T00:08:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.876208 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5139686878eb07d3abced92d0ccfe05be8907580ed75233b72fc2ad3415a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15cbe3fa75b69a7176803295e16a314a8bc73f5797c874124d54e76d47fa1064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:36Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.936750 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:36Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.956853 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfp2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"662dcbe7-7cf8-42d4-9a5e-345c32a49972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff3f3c69dcb28c7b0fd85f9e0c33b3339a1f6159dbf10bcf0efbdce4f7728c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a57d476b5b00110168878bd1c19c32ef122525c0259ac0a70de31c8220996d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4a57d476b5b00110168878bd1c19c32ef122525c0259ac0a70de31c8220996d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bb429b94a1286d8aac77cb77112c7163f7aa50a1c9cab997f8d088c100887c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0bb429b94a1286d8aac77cb77112c7163f7aa50a1c9cab997f8d088c100887c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a77485fd9e246f190724b86663e0dbe81146e1fcc51f468dd05f23c1db878fd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a77485fd9e246f190724b86663e0dbe81146e1fcc51f468dd05f23c1db878fd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adaa20d9a4898057802a8c9759ab3cd485963180bfc4f6b926bf6d2c886bcf27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adaa20d9a4898057802a8c9759ab3cd485963180bfc4f6b926bf6d2c886bcf27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://998fbb934f01c9ecd315f78e776199d4fd538106f8373f09b730d2df23ac9e3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://998fbb934f01c9ecd315f78e776199d4fd538106f8373f09b730d2df23ac9e3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46ebe18804ff07eb0e922a6cc1c39ff93875302d231222642b211e3527fe6373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46ebe18804ff07eb0e922a6cc1c39ff93875302d231222642b211e3527fe6373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfp2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:36Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.968365 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.968421 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.968439 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.968461 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.968478 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:36Z","lastTransitionTime":"2026-03-12T00:08:36Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.975068 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e483c2b-08f0-4e92-8e4a-b7281f30af3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d6839c99ab958ad1f40636c70e29bc285292fdbee1e910cdf7f8ab43ca7e8db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd9jv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82de8f2b633f1b6c295149d077bfd2b308f5acaa429acf411ed16f456536d0ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd9jv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4xwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:36Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:36 crc kubenswrapper[4948]: I0312 00:08:36.997501 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2efff53-8e12-41a1-bcb1-9a1d08c42a4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1cb328c8431b39dac735d49726a2eae40a367849b2c146bb2e49d02958f0e5a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ad13662849131222d5c8ba1041b81fa75bcb1fcbfa2e71c0c751f04b532a5b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8cd4faf47b25395ffd42d4bf5416ea29dcaa3ec380d6d6d6a9f13321d35195a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8028fd7b3e6d26bf30666425c843933b8fb87c3b2a258d510608ec718eb4fe82\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73834c8db23f6536034dce09e38b56665211238280bf9a4e1b13e74d81245a77\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T00:07:50Z\\\",\\\"message\\\":\\\"le observer\\\\nW0312 00:07:49.927227 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0312 00:07:49.927347 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 00:07:49.928000 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2511358960/tls.crt::/tmp/serving-cert-2511358960/tls.key\\\\\\\"\\\\nI0312 00:07:50.321455 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0312 00:07:50.324479 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0312 00:07:50.324511 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0312 00:07:50.324546 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0312 00:07:50.324556 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0312 00:07:50.331591 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0312 00:07:50.331609 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0312 00:07:50.331634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 00:07:50.331642 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 00:07:50.331651 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0312 00:07:50.331658 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0312 00:07:50.331664 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0312 00:07:50.331670 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0312 00:07:50.332917 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:07:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d56e32f9cb96ad103bab80d6a730b70d3387264a25a7a61ef1f0726e7290d6c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e42f2e94463ac8a3d7980a25122062da9ee87fcdc1f10e260ff9974b2e42359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e42f2e94463ac8a3d7980a25122062da9ee87fcdc1f10e260ff9974b2e42359\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:06:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:36Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:37 crc kubenswrapper[4948]: I0312 00:08:37.016895 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5a9ae9441886817563a5fc793bff7c9b8d77d406ceb1d1fb1eebd4307a1c64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:37Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:37 crc kubenswrapper[4948]: I0312 00:08:37.038408 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwpqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b7580c2-5c23-4c67-807a-ea97a3df9398\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d07c603619f858177d4b007e785208853abcfe2dc6c1af435309030fa409010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8lxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwpqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:37Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:37 crc kubenswrapper[4948]: I0312 00:08:37.053459 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-njc9c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0d6440c-4524-42ea-b18a-0ddc4662f672\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f4d0d9599ad62aaba35ba1414a806f3a163e29a3c34fd893ef5b14eae003591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7wn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-njc9c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:37Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:37 crc kubenswrapper[4948]: I0312 00:08:37.067042 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a124d35d-f697-49ae-8dac-ee32cd6fc5f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49c25293fe6ddb25b23fc250e3a5d94ce3f2c464f450c0da24dcff2febf15124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a57137a3fb48ab82c4324c6121b7dbcde8dd33e02cd7012ff42248cc5f164bec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccb319a44802931d007a0d9265f98bac6f23efee9a7fb82798718a315464e89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c756f1a5b429d17d95df4797725081682b66e9ef3e53a9deb9518e326e40325b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c756f1a5b429d17d95df4797725081682b66e9ef3e53a9deb9518e326e40325b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:06:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:37Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:37 crc kubenswrapper[4948]: I0312 00:08:37.070808 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:37 crc kubenswrapper[4948]: I0312 00:08:37.070838 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:37 crc kubenswrapper[4948]: I0312 00:08:37.070848 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:37 crc kubenswrapper[4948]: I0312 00:08:37.070862 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:37 crc kubenswrapper[4948]: I0312 00:08:37.070873 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:37Z","lastTransitionTime":"2026-03-12T00:08:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:37 crc kubenswrapper[4948]: I0312 00:08:37.079981 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9s5fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f9c7f1-6593-4d92-854c-e6c9964a0435\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b8833cbaf3809463e18bfebc9d5bada5fb26823831cc742a3a27177662f60e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xf4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9s5fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:37Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:37 crc kubenswrapper[4948]: I0312 00:08:37.090371 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xdd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b5233d6-d2b8-42dc-b881-18e439ecc442\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0ea55d993c72b6d50a4829ecfff4adae7d0f625a90e40545c9e63201cdda29c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjpwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb23ff4bc484524e306561c0c574e2b2eee39fc80efcd9f44f1924fe42df1a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjpwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4xdd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:37Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:37 crc kubenswrapper[4948]: I0312 00:08:37.173872 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:37 crc kubenswrapper[4948]: I0312 00:08:37.173910 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:37 crc kubenswrapper[4948]: I0312 00:08:37.173922 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:37 crc kubenswrapper[4948]: I0312 00:08:37.173941 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:37 crc kubenswrapper[4948]: I0312 00:08:37.173952 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:37Z","lastTransitionTime":"2026-03-12T00:08:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:37 crc kubenswrapper[4948]: I0312 00:08:37.277430 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:37 crc kubenswrapper[4948]: I0312 00:08:37.277492 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:37 crc kubenswrapper[4948]: I0312 00:08:37.277514 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:37 crc kubenswrapper[4948]: I0312 00:08:37.277540 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:37 crc kubenswrapper[4948]: I0312 00:08:37.277559 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:37Z","lastTransitionTime":"2026-03-12T00:08:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:37 crc kubenswrapper[4948]: I0312 00:08:37.388265 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:37 crc kubenswrapper[4948]: I0312 00:08:37.388312 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:37 crc kubenswrapper[4948]: I0312 00:08:37.388323 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:37 crc kubenswrapper[4948]: I0312 00:08:37.388337 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:37 crc kubenswrapper[4948]: I0312 00:08:37.388348 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:37Z","lastTransitionTime":"2026-03-12T00:08:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:37 crc kubenswrapper[4948]: I0312 00:08:37.490925 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:37 crc kubenswrapper[4948]: I0312 00:08:37.490998 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:37 crc kubenswrapper[4948]: I0312 00:08:37.491009 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:37 crc kubenswrapper[4948]: I0312 00:08:37.491022 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:37 crc kubenswrapper[4948]: I0312 00:08:37.491030 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:37Z","lastTransitionTime":"2026-03-12T00:08:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:37 crc kubenswrapper[4948]: I0312 00:08:37.594080 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:37 crc kubenswrapper[4948]: I0312 00:08:37.594146 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:37 crc kubenswrapper[4948]: I0312 00:08:37.594164 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:37 crc kubenswrapper[4948]: I0312 00:08:37.594190 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:37 crc kubenswrapper[4948]: I0312 00:08:37.594208 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:37Z","lastTransitionTime":"2026-03-12T00:08:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:37 crc kubenswrapper[4948]: I0312 00:08:37.697134 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:37 crc kubenswrapper[4948]: I0312 00:08:37.697202 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:37 crc kubenswrapper[4948]: I0312 00:08:37.697221 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:37 crc kubenswrapper[4948]: I0312 00:08:37.697244 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:37 crc kubenswrapper[4948]: I0312 00:08:37.697262 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:37Z","lastTransitionTime":"2026-03-12T00:08:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:37 crc kubenswrapper[4948]: I0312 00:08:37.800765 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:37 crc kubenswrapper[4948]: I0312 00:08:37.800828 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:37 crc kubenswrapper[4948]: I0312 00:08:37.800847 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:37 crc kubenswrapper[4948]: I0312 00:08:37.800874 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:37 crc kubenswrapper[4948]: I0312 00:08:37.800895 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:37Z","lastTransitionTime":"2026-03-12T00:08:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:37 crc kubenswrapper[4948]: I0312 00:08:37.903871 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:37 crc kubenswrapper[4948]: I0312 00:08:37.903920 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:37 crc kubenswrapper[4948]: I0312 00:08:37.903937 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:37 crc kubenswrapper[4948]: I0312 00:08:37.903961 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:37 crc kubenswrapper[4948]: I0312 00:08:37.903978 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:37Z","lastTransitionTime":"2026-03-12T00:08:37Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:38 crc kubenswrapper[4948]: I0312 00:08:38.006902 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:38 crc kubenswrapper[4948]: I0312 00:08:38.006964 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:38 crc kubenswrapper[4948]: I0312 00:08:38.006981 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:38 crc kubenswrapper[4948]: I0312 00:08:38.007007 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:38 crc kubenswrapper[4948]: I0312 00:08:38.007029 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:38Z","lastTransitionTime":"2026-03-12T00:08:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:38 crc kubenswrapper[4948]: I0312 00:08:38.110069 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:38 crc kubenswrapper[4948]: I0312 00:08:38.111092 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:38 crc kubenswrapper[4948]: I0312 00:08:38.111269 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:38 crc kubenswrapper[4948]: I0312 00:08:38.111570 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:38 crc kubenswrapper[4948]: I0312 00:08:38.111778 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:38Z","lastTransitionTime":"2026-03-12T00:08:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:38 crc kubenswrapper[4948]: I0312 00:08:38.214649 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:38 crc kubenswrapper[4948]: I0312 00:08:38.214741 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:38 crc kubenswrapper[4948]: I0312 00:08:38.214759 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:38 crc kubenswrapper[4948]: I0312 00:08:38.214782 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:38 crc kubenswrapper[4948]: I0312 00:08:38.214798 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:38Z","lastTransitionTime":"2026-03-12T00:08:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:38 crc kubenswrapper[4948]: I0312 00:08:38.312488 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:08:38 crc kubenswrapper[4948]: I0312 00:08:38.312497 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:08:38 crc kubenswrapper[4948]: E0312 00:08:38.313034 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 00:08:38 crc kubenswrapper[4948]: I0312 00:08:38.312546 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:08:38 crc kubenswrapper[4948]: I0312 00:08:38.312520 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:08:38 crc kubenswrapper[4948]: E0312 00:08:38.313142 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 00:08:38 crc kubenswrapper[4948]: E0312 00:08:38.313287 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mqmtd" podUID="be985ff1-de8a-431f-a230-67894ebe52d5" Mar 12 00:08:38 crc kubenswrapper[4948]: E0312 00:08:38.312912 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 00:08:38 crc kubenswrapper[4948]: I0312 00:08:38.318056 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:38 crc kubenswrapper[4948]: I0312 00:08:38.318110 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:38 crc kubenswrapper[4948]: I0312 00:08:38.318132 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:38 crc kubenswrapper[4948]: I0312 00:08:38.318160 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:38 crc kubenswrapper[4948]: I0312 00:08:38.318181 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:38Z","lastTransitionTime":"2026-03-12T00:08:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:38 crc kubenswrapper[4948]: I0312 00:08:38.421597 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:38 crc kubenswrapper[4948]: I0312 00:08:38.421651 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:38 crc kubenswrapper[4948]: I0312 00:08:38.421670 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:38 crc kubenswrapper[4948]: I0312 00:08:38.421694 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:38 crc kubenswrapper[4948]: I0312 00:08:38.421714 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:38Z","lastTransitionTime":"2026-03-12T00:08:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:38 crc kubenswrapper[4948]: I0312 00:08:38.524870 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:38 crc kubenswrapper[4948]: I0312 00:08:38.524926 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:38 crc kubenswrapper[4948]: I0312 00:08:38.524943 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:38 crc kubenswrapper[4948]: I0312 00:08:38.524963 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:38 crc kubenswrapper[4948]: I0312 00:08:38.524978 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:38Z","lastTransitionTime":"2026-03-12T00:08:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:38 crc kubenswrapper[4948]: I0312 00:08:38.632402 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:38 crc kubenswrapper[4948]: I0312 00:08:38.632477 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:38 crc kubenswrapper[4948]: I0312 00:08:38.632500 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:38 crc kubenswrapper[4948]: I0312 00:08:38.632530 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:38 crc kubenswrapper[4948]: I0312 00:08:38.632561 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:38Z","lastTransitionTime":"2026-03-12T00:08:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:38 crc kubenswrapper[4948]: I0312 00:08:38.735276 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:38 crc kubenswrapper[4948]: I0312 00:08:38.735367 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:38 crc kubenswrapper[4948]: I0312 00:08:38.735384 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:38 crc kubenswrapper[4948]: I0312 00:08:38.735407 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:38 crc kubenswrapper[4948]: I0312 00:08:38.735423 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:38Z","lastTransitionTime":"2026-03-12T00:08:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:38 crc kubenswrapper[4948]: I0312 00:08:38.838509 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:38 crc kubenswrapper[4948]: I0312 00:08:38.838573 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:38 crc kubenswrapper[4948]: I0312 00:08:38.838597 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:38 crc kubenswrapper[4948]: I0312 00:08:38.838624 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:38 crc kubenswrapper[4948]: I0312 00:08:38.838646 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:38Z","lastTransitionTime":"2026-03-12T00:08:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:38 crc kubenswrapper[4948]: I0312 00:08:38.941462 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:38 crc kubenswrapper[4948]: I0312 00:08:38.941510 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:38 crc kubenswrapper[4948]: I0312 00:08:38.941527 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:38 crc kubenswrapper[4948]: I0312 00:08:38.941549 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:38 crc kubenswrapper[4948]: I0312 00:08:38.941567 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:38Z","lastTransitionTime":"2026-03-12T00:08:38Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:39 crc kubenswrapper[4948]: I0312 00:08:39.044582 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:39 crc kubenswrapper[4948]: I0312 00:08:39.044651 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:39 crc kubenswrapper[4948]: I0312 00:08:39.044677 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:39 crc kubenswrapper[4948]: I0312 00:08:39.044708 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:39 crc kubenswrapper[4948]: I0312 00:08:39.044729 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:39Z","lastTransitionTime":"2026-03-12T00:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:39 crc kubenswrapper[4948]: I0312 00:08:39.147682 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:39 crc kubenswrapper[4948]: I0312 00:08:39.147753 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:39 crc kubenswrapper[4948]: I0312 00:08:39.147773 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:39 crc kubenswrapper[4948]: I0312 00:08:39.147801 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:39 crc kubenswrapper[4948]: I0312 00:08:39.147819 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:39Z","lastTransitionTime":"2026-03-12T00:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:39 crc kubenswrapper[4948]: I0312 00:08:39.250078 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:39 crc kubenswrapper[4948]: I0312 00:08:39.250152 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:39 crc kubenswrapper[4948]: I0312 00:08:39.250175 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:39 crc kubenswrapper[4948]: I0312 00:08:39.250205 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:39 crc kubenswrapper[4948]: I0312 00:08:39.250229 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:39Z","lastTransitionTime":"2026-03-12T00:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:39 crc kubenswrapper[4948]: I0312 00:08:39.353943 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:39 crc kubenswrapper[4948]: I0312 00:08:39.354010 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:39 crc kubenswrapper[4948]: I0312 00:08:39.354031 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:39 crc kubenswrapper[4948]: I0312 00:08:39.354057 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:39 crc kubenswrapper[4948]: I0312 00:08:39.354075 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:39Z","lastTransitionTime":"2026-03-12T00:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:39 crc kubenswrapper[4948]: I0312 00:08:39.457085 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:39 crc kubenswrapper[4948]: I0312 00:08:39.457149 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:39 crc kubenswrapper[4948]: I0312 00:08:39.457167 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:39 crc kubenswrapper[4948]: I0312 00:08:39.457194 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:39 crc kubenswrapper[4948]: I0312 00:08:39.457210 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:39Z","lastTransitionTime":"2026-03-12T00:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:39 crc kubenswrapper[4948]: I0312 00:08:39.560044 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:39 crc kubenswrapper[4948]: I0312 00:08:39.560121 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:39 crc kubenswrapper[4948]: I0312 00:08:39.560145 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:39 crc kubenswrapper[4948]: I0312 00:08:39.560172 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:39 crc kubenswrapper[4948]: I0312 00:08:39.560192 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:39Z","lastTransitionTime":"2026-03-12T00:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:39 crc kubenswrapper[4948]: I0312 00:08:39.662846 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:39 crc kubenswrapper[4948]: I0312 00:08:39.662916 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:39 crc kubenswrapper[4948]: I0312 00:08:39.662933 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:39 crc kubenswrapper[4948]: I0312 00:08:39.662956 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:39 crc kubenswrapper[4948]: I0312 00:08:39.662974 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:39Z","lastTransitionTime":"2026-03-12T00:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:39 crc kubenswrapper[4948]: I0312 00:08:39.766268 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:39 crc kubenswrapper[4948]: I0312 00:08:39.766357 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:39 crc kubenswrapper[4948]: I0312 00:08:39.766376 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:39 crc kubenswrapper[4948]: I0312 00:08:39.766403 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:39 crc kubenswrapper[4948]: I0312 00:08:39.766422 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:39Z","lastTransitionTime":"2026-03-12T00:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:39 crc kubenswrapper[4948]: I0312 00:08:39.869048 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:39 crc kubenswrapper[4948]: I0312 00:08:39.869111 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:39 crc kubenswrapper[4948]: I0312 00:08:39.869128 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:39 crc kubenswrapper[4948]: I0312 00:08:39.869152 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:39 crc kubenswrapper[4948]: I0312 00:08:39.869170 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:39Z","lastTransitionTime":"2026-03-12T00:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:39 crc kubenswrapper[4948]: I0312 00:08:39.972624 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:39 crc kubenswrapper[4948]: I0312 00:08:39.972701 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:39 crc kubenswrapper[4948]: I0312 00:08:39.972729 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:39 crc kubenswrapper[4948]: I0312 00:08:39.972758 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:39 crc kubenswrapper[4948]: I0312 00:08:39.972781 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:39Z","lastTransitionTime":"2026-03-12T00:08:39Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:40 crc kubenswrapper[4948]: I0312 00:08:40.076056 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:40 crc kubenswrapper[4948]: I0312 00:08:40.076111 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:40 crc kubenswrapper[4948]: I0312 00:08:40.076134 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:40 crc kubenswrapper[4948]: I0312 00:08:40.076161 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:40 crc kubenswrapper[4948]: I0312 00:08:40.076184 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:40Z","lastTransitionTime":"2026-03-12T00:08:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:40 crc kubenswrapper[4948]: I0312 00:08:40.179980 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:40 crc kubenswrapper[4948]: I0312 00:08:40.180039 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:40 crc kubenswrapper[4948]: I0312 00:08:40.180057 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:40 crc kubenswrapper[4948]: I0312 00:08:40.180081 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:40 crc kubenswrapper[4948]: I0312 00:08:40.180102 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:40Z","lastTransitionTime":"2026-03-12T00:08:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:40 crc kubenswrapper[4948]: I0312 00:08:40.282659 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:40 crc kubenswrapper[4948]: I0312 00:08:40.282711 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:40 crc kubenswrapper[4948]: I0312 00:08:40.282727 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:40 crc kubenswrapper[4948]: I0312 00:08:40.282752 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:40 crc kubenswrapper[4948]: I0312 00:08:40.282770 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:40Z","lastTransitionTime":"2026-03-12T00:08:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:40 crc kubenswrapper[4948]: I0312 00:08:40.312275 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:08:40 crc kubenswrapper[4948]: I0312 00:08:40.312355 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:08:40 crc kubenswrapper[4948]: I0312 00:08:40.312395 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:08:40 crc kubenswrapper[4948]: I0312 00:08:40.312355 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:08:40 crc kubenswrapper[4948]: E0312 00:08:40.312494 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 00:08:40 crc kubenswrapper[4948]: E0312 00:08:40.312650 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mqmtd" podUID="be985ff1-de8a-431f-a230-67894ebe52d5" Mar 12 00:08:40 crc kubenswrapper[4948]: E0312 00:08:40.312700 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 00:08:40 crc kubenswrapper[4948]: E0312 00:08:40.312807 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 00:08:40 crc kubenswrapper[4948]: I0312 00:08:40.385661 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:40 crc kubenswrapper[4948]: I0312 00:08:40.385826 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:40 crc kubenswrapper[4948]: I0312 00:08:40.385857 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:40 crc kubenswrapper[4948]: I0312 00:08:40.385891 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:40 crc kubenswrapper[4948]: I0312 00:08:40.385916 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:40Z","lastTransitionTime":"2026-03-12T00:08:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:40 crc kubenswrapper[4948]: I0312 00:08:40.488694 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:40 crc kubenswrapper[4948]: I0312 00:08:40.488755 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:40 crc kubenswrapper[4948]: I0312 00:08:40.488773 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:40 crc kubenswrapper[4948]: I0312 00:08:40.488799 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:40 crc kubenswrapper[4948]: I0312 00:08:40.488818 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:40Z","lastTransitionTime":"2026-03-12T00:08:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:40 crc kubenswrapper[4948]: I0312 00:08:40.591584 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:40 crc kubenswrapper[4948]: I0312 00:08:40.591642 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:40 crc kubenswrapper[4948]: I0312 00:08:40.591659 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:40 crc kubenswrapper[4948]: I0312 00:08:40.591682 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:40 crc kubenswrapper[4948]: I0312 00:08:40.591699 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:40Z","lastTransitionTime":"2026-03-12T00:08:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:40 crc kubenswrapper[4948]: I0312 00:08:40.695287 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:40 crc kubenswrapper[4948]: I0312 00:08:40.695381 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:40 crc kubenswrapper[4948]: I0312 00:08:40.695405 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:40 crc kubenswrapper[4948]: I0312 00:08:40.695438 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:40 crc kubenswrapper[4948]: I0312 00:08:40.695461 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:40Z","lastTransitionTime":"2026-03-12T00:08:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:40 crc kubenswrapper[4948]: I0312 00:08:40.798092 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:40 crc kubenswrapper[4948]: I0312 00:08:40.798146 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:40 crc kubenswrapper[4948]: I0312 00:08:40.798175 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:40 crc kubenswrapper[4948]: I0312 00:08:40.798220 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:40 crc kubenswrapper[4948]: I0312 00:08:40.798244 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:40Z","lastTransitionTime":"2026-03-12T00:08:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:40 crc kubenswrapper[4948]: I0312 00:08:40.901255 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:40 crc kubenswrapper[4948]: I0312 00:08:40.901343 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:40 crc kubenswrapper[4948]: I0312 00:08:40.901369 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:40 crc kubenswrapper[4948]: I0312 00:08:40.901390 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:40 crc kubenswrapper[4948]: I0312 00:08:40.901406 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:40Z","lastTransitionTime":"2026-03-12T00:08:40Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:41 crc kubenswrapper[4948]: I0312 00:08:41.004626 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:41 crc kubenswrapper[4948]: I0312 00:08:41.004694 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:41 crc kubenswrapper[4948]: I0312 00:08:41.004718 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:41 crc kubenswrapper[4948]: I0312 00:08:41.004745 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:41 crc kubenswrapper[4948]: I0312 00:08:41.004768 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:41Z","lastTransitionTime":"2026-03-12T00:08:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:41 crc kubenswrapper[4948]: I0312 00:08:41.107477 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:41 crc kubenswrapper[4948]: I0312 00:08:41.107527 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:41 crc kubenswrapper[4948]: I0312 00:08:41.107544 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:41 crc kubenswrapper[4948]: I0312 00:08:41.107566 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:41 crc kubenswrapper[4948]: I0312 00:08:41.107582 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:41Z","lastTransitionTime":"2026-03-12T00:08:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:41 crc kubenswrapper[4948]: I0312 00:08:41.210726 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:41 crc kubenswrapper[4948]: I0312 00:08:41.210833 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:41 crc kubenswrapper[4948]: I0312 00:08:41.210900 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:41 crc kubenswrapper[4948]: I0312 00:08:41.210927 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:41 crc kubenswrapper[4948]: I0312 00:08:41.210945 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:41Z","lastTransitionTime":"2026-03-12T00:08:41Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:41 crc kubenswrapper[4948]: E0312 00:08:41.312293 4948 kubelet_node_status.go:497] "Node not becoming ready in time after startup" Mar 12 00:08:41 crc kubenswrapper[4948]: I0312 00:08:41.315643 4948 scope.go:117] "RemoveContainer" containerID="588e8a26bafbb970547438b573ab34f111ef18d8134df2b3fff513c9cca9a516" Mar 12 00:08:41 crc kubenswrapper[4948]: I0312 00:08:41.335059 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5a9ae9441886817563a5fc793bff7c9b8d77d406ceb1d1fb1eebd4307a1c64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:41Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:41 crc kubenswrapper[4948]: I0312 00:08:41.355898 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwpqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b7580c2-5c23-4c67-807a-ea97a3df9398\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d07c603619f858177d4b007e785208853abcfe2dc6c1af435309030fa409010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8lxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwpqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:41Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:41 crc kubenswrapper[4948]: I0312 00:08:41.374178 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-njc9c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0d6440c-4524-42ea-b18a-0ddc4662f672\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f4d0d9599ad62aaba35ba1414a806f3a163e29a3c34fd893ef5b14eae003591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7wn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-njc9c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:41Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:41 crc kubenswrapper[4948]: I0312 00:08:41.396964 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2efff53-8e12-41a1-bcb1-9a1d08c42a4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1cb328c8431b39dac735d49726a2eae40a367849b2c146bb2e49d02958f0e5a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ad13662849131222d5c8ba1041b81fa75bcb1fcbfa2e71c0c751f04b532a5b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8cd4faf47b25395ffd42d4bf5416ea29dcaa3ec380d6d6d6a9f13321d35195a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8028fd7b3e6d26bf30666425c843933b8fb87c3b2a258d510608ec718eb4fe82\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73834c8db23f6536034dce09e38b56665211238280bf9a4e1b13e74d81245a77\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T00:07:50Z\\\",\\\"message\\\":\\\"le observer\\\\nW0312 00:07:49.927227 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0312 00:07:49.927347 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 00:07:49.928000 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2511358960/tls.crt::/tmp/serving-cert-2511358960/tls.key\\\\\\\"\\\\nI0312 00:07:50.321455 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0312 00:07:50.324479 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0312 00:07:50.324511 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0312 00:07:50.324546 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0312 00:07:50.324556 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0312 00:07:50.331591 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0312 00:07:50.331609 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0312 00:07:50.331634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 00:07:50.331642 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 00:07:50.331651 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0312 00:07:50.331658 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0312 00:07:50.331664 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0312 00:07:50.331670 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0312 00:07:50.332917 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:07:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d56e32f9cb96ad103bab80d6a730b70d3387264a25a7a61ef1f0726e7290d6c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e42f2e94463ac8a3d7980a25122062da9ee87fcdc1f10e260ff9974b2e42359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e42f2e94463ac8a3d7980a25122062da9ee87fcdc1f10e260ff9974b2e42359\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:06:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:41Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:41 crc kubenswrapper[4948]: I0312 00:08:41.415788 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xdd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b5233d6-d2b8-42dc-b881-18e439ecc442\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0ea55d993c72b6d50a4829ecfff4adae7d0f625a90e40545c9e63201cdda29c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjpwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb23ff4bc484524e306561c0c574e2b2eee39fc80efcd9f44f1924fe42df1a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjpwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4xdd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:41Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:41 crc kubenswrapper[4948]: E0312 00:08:41.435163 4948 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 12 00:08:41 crc kubenswrapper[4948]: I0312 00:08:41.436961 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a124d35d-f697-49ae-8dac-ee32cd6fc5f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49c25293fe6ddb25b23fc250e3a5d94ce3f2c464f450c0da24dcff2febf15124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a57137a3fb48ab82c4324c6121b7dbcde8dd33e02cd7012ff42248cc5f164bec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccb319a44802931d007a0d9265f98bac6f23efee9a7fb82798718a315464e89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c756f1a5b429d17d95df4797725081682b66e9ef3e53a9deb9518e326e40325b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c756f1a5b429d17d95df4797725081682b66e9ef3e53a9deb9518e326e40325b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:06:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:41Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:41 crc kubenswrapper[4948]: I0312 00:08:41.454119 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9s5fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f9c7f1-6593-4d92-854c-e6c9964a0435\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b8833cbaf3809463e18bfebc9d5bada5fb26823831cc742a3a27177662f60e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xf4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9s5fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:41Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:41 crc kubenswrapper[4948]: I0312 00:08:41.473805 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8d58ee45d75284e030700fa526b2c39197a26f6a391a7417f25a026f356b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:41Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:41 crc kubenswrapper[4948]: I0312 00:08:41.493702 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mqmtd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be985ff1-de8a-431f-a230-67894ebe52d5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmsz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmsz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mqmtd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:41Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:41 crc kubenswrapper[4948]: I0312 00:08:41.529501 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15117613-9fad-48c7-98c4-a2d84502ded9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4036d7131db00afd8eab8e093626b688297dc776b5af80c3c38b51da7124085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84af4cdd213d90636a889a8a8a3aaca3dee1e2677368efa0ffc768940c2cc994\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e481ace9e84b6aa66cbdda3ee8425c4120a665915d2c70ae4faf60bba6205e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9eadefd0babc119f776d48c9d2c9e4bced58db685af9a8e97f25e6a9a2e3c89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9d5e77753fda70350353daa0967257dd4929dba4440a95495ab4c2a0659aef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cad943edb9a8d445b33a6e70d85b475c1ccba460a6056bccae94fe60fd6b73f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://588e8a26bafbb970547438b573ab34f111ef18d8134df2b3fff513c9cca9a516\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://588e8a26bafbb970547438b573ab34f111ef18d8134df2b3fff513c9cca9a516\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T00:08:25Z\\\",\\\"message\\\":\\\"umn _uuid == {f6d604c1-9711-4e25-be6c-79ec28bbad1b}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0312 00:08:25.737942 6942 obj_retry.go:551] Creating *factory.egressNode crc took: 2.187304ms\\\\nI0312 00:08:25.737970 6942 factory.go:1336] Added *v1.Node event handler 7\\\\nI0312 00:08:25.738003 6942 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI0312 00:08:25.738210 6942 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0312 00:08:25.738274 6942 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0312 00:08:25.738325 6942 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0312 00:08:25.738333 6942 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0312 00:08:25.738357 6942 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0312 00:08:25.738373 6942 factory.go:656] Stopping watch factory\\\\nI0312 00:08:25.738386 6942 ovnkube.go:599] Stopped ovnkube\\\\nI0312 00:08:25.738410 6942 handler.go:208] Removed *v1.Node event handler 2\\\\nI0312 00:08:25.738414 6942 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0312 00:08:25.738430 6942 handler.go:208] Removed *v1.Node event handler 7\\\\nI0312 00:08:25.738442 6942 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nF0312 00:08:25.738550 6942 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:24Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":1,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 10s restarting failed container=ovnkube-controller pod=ovnkube-node-f7mp5_openshift-ovn-kubernetes(15117613-9fad-48c7-98c4-a2d84502ded9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46fa6140f4f4bd32e298fa36f6d7af6890029f9bc3080d2b4d94be8307346de8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7mp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:41Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:41 crc kubenswrapper[4948]: I0312 00:08:41.549265 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:41Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:41 crc kubenswrapper[4948]: I0312 00:08:41.570222 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:41Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:41 crc kubenswrapper[4948]: I0312 00:08:41.588259 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5139686878eb07d3abced92d0ccfe05be8907580ed75233b72fc2ad3415a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15cbe3fa75b69a7176803295e16a314a8bc73f5797c874124d54e76d47fa1064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:41Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:41 crc kubenswrapper[4948]: I0312 00:08:41.606488 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:41Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:41 crc kubenswrapper[4948]: I0312 00:08:41.627286 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfp2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"662dcbe7-7cf8-42d4-9a5e-345c32a49972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff3f3c69dcb28c7b0fd85f9e0c33b3339a1f6159dbf10bcf0efbdce4f7728c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a57d476b5b00110168878bd1c19c32ef122525c0259ac0a70de31c8220996d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4a57d476b5b00110168878bd1c19c32ef122525c0259ac0a70de31c8220996d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bb429b94a1286d8aac77cb77112c7163f7aa50a1c9cab997f8d088c100887c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0bb429b94a1286d8aac77cb77112c7163f7aa50a1c9cab997f8d088c100887c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a77485fd9e246f190724b86663e0dbe81146e1fcc51f468dd05f23c1db878fd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a77485fd9e246f190724b86663e0dbe81146e1fcc51f468dd05f23c1db878fd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adaa20d9a4898057802a8c9759ab3cd485963180bfc4f6b926bf6d2c886bcf27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adaa20d9a4898057802a8c9759ab3cd485963180bfc4f6b926bf6d2c886bcf27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://998fbb934f01c9ecd315f78e776199d4fd538106f8373f09b730d2df23ac9e3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://998fbb934f01c9ecd315f78e776199d4fd538106f8373f09b730d2df23ac9e3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46ebe18804ff07eb0e922a6cc1c39ff93875302d231222642b211e3527fe6373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46ebe18804ff07eb0e922a6cc1c39ff93875302d231222642b211e3527fe6373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfp2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:41Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:41 crc kubenswrapper[4948]: I0312 00:08:41.641562 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e483c2b-08f0-4e92-8e4a-b7281f30af3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d6839c99ab958ad1f40636c70e29bc285292fdbee1e910cdf7f8ab43ca7e8db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd9jv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82de8f2b633f1b6c295149d077bfd2b308f5acaa429acf411ed16f456536d0ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd9jv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4xwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:41Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:41 crc kubenswrapper[4948]: I0312 00:08:41.652325 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c835bb7a-4e06-4047-9fe6-e0c605b36bc6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc9760bd6ed0d092081fdd2ad33dda2d67f1305531640a35b9dce84b5d86a813\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f580535b16c31fb0ebadbdd9b7851ba2dce2f7d52bcad8c9cacc197f5e7dfad4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T00:07:08Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0312 00:06:43.492131 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0312 00:06:43.494442 1 observer_polling.go:159] Starting file observer\\\\nI0312 00:06:43.528056 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0312 00:06:43.536640 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0312 00:07:08.239474 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0312 00:07:08.239540 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb45482a404d6696cfbc3919997e53146b7b9dd1536056a55374574ae01ff193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://663bca87e44109c1d2bb59964e567d8b78fe2e9de39bcec0557cb43e39143add\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bf5b5d40a0cc76725ae2cf453089b31e75c304cf6037ef06b8ca8f7cff35a5a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:06:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:41Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:41 crc kubenswrapper[4948]: I0312 00:08:41.747233 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f7mp5_15117613-9fad-48c7-98c4-a2d84502ded9/ovnkube-controller/1.log" Mar 12 00:08:41 crc kubenswrapper[4948]: I0312 00:08:41.751780 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" event={"ID":"15117613-9fad-48c7-98c4-a2d84502ded9","Type":"ContainerStarted","Data":"a18330e68c4cbdf0380d4f4845bed125a80893c948f89c674632af8eb1ffb5f4"} Mar 12 00:08:41 crc kubenswrapper[4948]: I0312 00:08:41.753141 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:08:41 crc kubenswrapper[4948]: I0312 00:08:41.777503 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2efff53-8e12-41a1-bcb1-9a1d08c42a4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1cb328c8431b39dac735d49726a2eae40a367849b2c146bb2e49d02958f0e5a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ad13662849131222d5c8ba1041b81fa75bcb1fcbfa2e71c0c751f04b532a5b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8cd4faf47b25395ffd42d4bf5416ea29dcaa3ec380d6d6d6a9f13321d35195a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8028fd7b3e6d26bf30666425c843933b8fb87c3b2a258d510608ec718eb4fe82\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73834c8db23f6536034dce09e38b56665211238280bf9a4e1b13e74d81245a77\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T00:07:50Z\\\",\\\"message\\\":\\\"le observer\\\\nW0312 00:07:49.927227 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0312 00:07:49.927347 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 00:07:49.928000 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2511358960/tls.crt::/tmp/serving-cert-2511358960/tls.key\\\\\\\"\\\\nI0312 00:07:50.321455 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0312 00:07:50.324479 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0312 00:07:50.324511 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0312 00:07:50.324546 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0312 00:07:50.324556 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0312 00:07:50.331591 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0312 00:07:50.331609 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0312 00:07:50.331634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 00:07:50.331642 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 00:07:50.331651 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0312 00:07:50.331658 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0312 00:07:50.331664 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0312 00:07:50.331670 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0312 00:07:50.332917 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:07:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d56e32f9cb96ad103bab80d6a730b70d3387264a25a7a61ef1f0726e7290d6c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e42f2e94463ac8a3d7980a25122062da9ee87fcdc1f10e260ff9974b2e42359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e42f2e94463ac8a3d7980a25122062da9ee87fcdc1f10e260ff9974b2e42359\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:06:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:41Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:41 crc kubenswrapper[4948]: I0312 00:08:41.800091 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5a9ae9441886817563a5fc793bff7c9b8d77d406ceb1d1fb1eebd4307a1c64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:41Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:41 crc kubenswrapper[4948]: I0312 00:08:41.825847 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwpqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b7580c2-5c23-4c67-807a-ea97a3df9398\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d07c603619f858177d4b007e785208853abcfe2dc6c1af435309030fa409010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8lxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwpqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:41Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:41 crc kubenswrapper[4948]: I0312 00:08:41.843503 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-njc9c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0d6440c-4524-42ea-b18a-0ddc4662f672\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f4d0d9599ad62aaba35ba1414a806f3a163e29a3c34fd893ef5b14eae003591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7wn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-njc9c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:41Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:41 crc kubenswrapper[4948]: I0312 00:08:41.861587 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a124d35d-f697-49ae-8dac-ee32cd6fc5f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49c25293fe6ddb25b23fc250e3a5d94ce3f2c464f450c0da24dcff2febf15124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a57137a3fb48ab82c4324c6121b7dbcde8dd33e02cd7012ff42248cc5f164bec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccb319a44802931d007a0d9265f98bac6f23efee9a7fb82798718a315464e89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c756f1a5b429d17d95df4797725081682b66e9ef3e53a9deb9518e326e40325b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c756f1a5b429d17d95df4797725081682b66e9ef3e53a9deb9518e326e40325b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:06:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:41Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:41 crc kubenswrapper[4948]: I0312 00:08:41.883390 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9s5fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f9c7f1-6593-4d92-854c-e6c9964a0435\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b8833cbaf3809463e18bfebc9d5bada5fb26823831cc742a3a27177662f60e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xf4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9s5fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:41Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:41 crc kubenswrapper[4948]: I0312 00:08:41.944866 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xdd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b5233d6-d2b8-42dc-b881-18e439ecc442\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0ea55d993c72b6d50a4829ecfff4adae7d0f625a90e40545c9e63201cdda29c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjpwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb23ff4bc484524e306561c0c574e2b2eee39fc80efcd9f44f1924fe42df1a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjpwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4xdd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:41Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:41 crc kubenswrapper[4948]: I0312 00:08:41.961985 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:41Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:41 crc kubenswrapper[4948]: I0312 00:08:41.975644 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:41Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:41 crc kubenswrapper[4948]: I0312 00:08:41.986166 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8d58ee45d75284e030700fa526b2c39197a26f6a391a7417f25a026f356b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:41Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:41 crc kubenswrapper[4948]: I0312 00:08:41.997139 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mqmtd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be985ff1-de8a-431f-a230-67894ebe52d5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmsz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmsz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mqmtd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:41Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:42 crc kubenswrapper[4948]: I0312 00:08:42.023116 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15117613-9fad-48c7-98c4-a2d84502ded9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4036d7131db00afd8eab8e093626b688297dc776b5af80c3c38b51da7124085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84af4cdd213d90636a889a8a8a3aaca3dee1e2677368efa0ffc768940c2cc994\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e481ace9e84b6aa66cbdda3ee8425c4120a665915d2c70ae4faf60bba6205e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9eadefd0babc119f776d48c9d2c9e4bced58db685af9a8e97f25e6a9a2e3c89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9d5e77753fda70350353daa0967257dd4929dba4440a95495ab4c2a0659aef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cad943edb9a8d445b33a6e70d85b475c1ccba460a6056bccae94fe60fd6b73f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a18330e68c4cbdf0380d4f4845bed125a80893c948f89c674632af8eb1ffb5f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://588e8a26bafbb970547438b573ab34f111ef18d8134df2b3fff513c9cca9a516\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T00:08:25Z\\\",\\\"message\\\":\\\"umn _uuid == {f6d604c1-9711-4e25-be6c-79ec28bbad1b}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0312 00:08:25.737942 6942 obj_retry.go:551] Creating *factory.egressNode crc took: 2.187304ms\\\\nI0312 00:08:25.737970 6942 factory.go:1336] Added *v1.Node event handler 7\\\\nI0312 00:08:25.738003 6942 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI0312 00:08:25.738210 6942 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0312 00:08:25.738274 6942 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0312 00:08:25.738325 6942 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0312 00:08:25.738333 6942 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0312 00:08:25.738357 6942 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0312 00:08:25.738373 6942 factory.go:656] Stopping watch factory\\\\nI0312 00:08:25.738386 6942 ovnkube.go:599] Stopped ovnkube\\\\nI0312 00:08:25.738410 6942 handler.go:208] Removed *v1.Node event handler 2\\\\nI0312 00:08:25.738414 6942 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0312 00:08:25.738430 6942 handler.go:208] Removed *v1.Node event handler 7\\\\nI0312 00:08:25.738442 6942 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nF0312 00:08:25.738550 6942 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:24Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46fa6140f4f4bd32e298fa36f6d7af6890029f9bc3080d2b4d94be8307346de8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7mp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:42Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:42 crc kubenswrapper[4948]: I0312 00:08:42.039112 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c835bb7a-4e06-4047-9fe6-e0c605b36bc6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc9760bd6ed0d092081fdd2ad33dda2d67f1305531640a35b9dce84b5d86a813\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f580535b16c31fb0ebadbdd9b7851ba2dce2f7d52bcad8c9cacc197f5e7dfad4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T00:07:08Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0312 00:06:43.492131 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0312 00:06:43.494442 1 observer_polling.go:159] Starting file observer\\\\nI0312 00:06:43.528056 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0312 00:06:43.536640 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0312 00:07:08.239474 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0312 00:07:08.239540 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb45482a404d6696cfbc3919997e53146b7b9dd1536056a55374574ae01ff193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://663bca87e44109c1d2bb59964e567d8b78fe2e9de39bcec0557cb43e39143add\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bf5b5d40a0cc76725ae2cf453089b31e75c304cf6037ef06b8ca8f7cff35a5a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:06:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:42Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:42 crc kubenswrapper[4948]: I0312 00:08:42.059260 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5139686878eb07d3abced92d0ccfe05be8907580ed75233b72fc2ad3415a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15cbe3fa75b69a7176803295e16a314a8bc73f5797c874124d54e76d47fa1064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:42Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:42 crc kubenswrapper[4948]: I0312 00:08:42.075808 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:42Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:42 crc kubenswrapper[4948]: I0312 00:08:42.103116 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfp2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"662dcbe7-7cf8-42d4-9a5e-345c32a49972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff3f3c69dcb28c7b0fd85f9e0c33b3339a1f6159dbf10bcf0efbdce4f7728c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a57d476b5b00110168878bd1c19c32ef122525c0259ac0a70de31c8220996d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4a57d476b5b00110168878bd1c19c32ef122525c0259ac0a70de31c8220996d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bb429b94a1286d8aac77cb77112c7163f7aa50a1c9cab997f8d088c100887c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0bb429b94a1286d8aac77cb77112c7163f7aa50a1c9cab997f8d088c100887c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a77485fd9e246f190724b86663e0dbe81146e1fcc51f468dd05f23c1db878fd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a77485fd9e246f190724b86663e0dbe81146e1fcc51f468dd05f23c1db878fd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adaa20d9a4898057802a8c9759ab3cd485963180bfc4f6b926bf6d2c886bcf27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adaa20d9a4898057802a8c9759ab3cd485963180bfc4f6b926bf6d2c886bcf27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://998fbb934f01c9ecd315f78e776199d4fd538106f8373f09b730d2df23ac9e3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://998fbb934f01c9ecd315f78e776199d4fd538106f8373f09b730d2df23ac9e3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46ebe18804ff07eb0e922a6cc1c39ff93875302d231222642b211e3527fe6373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46ebe18804ff07eb0e922a6cc1c39ff93875302d231222642b211e3527fe6373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfp2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:42Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:42 crc kubenswrapper[4948]: I0312 00:08:42.118670 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e483c2b-08f0-4e92-8e4a-b7281f30af3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d6839c99ab958ad1f40636c70e29bc285292fdbee1e910cdf7f8ab43ca7e8db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd9jv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82de8f2b633f1b6c295149d077bfd2b308f5acaa429acf411ed16f456536d0ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd9jv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4xwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:42Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:42 crc kubenswrapper[4948]: I0312 00:08:42.156736 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:42 crc kubenswrapper[4948]: I0312 00:08:42.156910 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:42 crc kubenswrapper[4948]: I0312 00:08:42.156996 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:42 crc kubenswrapper[4948]: I0312 00:08:42.157085 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:42 crc kubenswrapper[4948]: I0312 00:08:42.157166 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:42Z","lastTransitionTime":"2026-03-12T00:08:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:42 crc kubenswrapper[4948]: E0312 00:08:42.170866 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17825588-29cf-4059-bab7-1b042b9bf1f3\\\",\\\"systemUUID\\\":\\\"b18e4bf2-d37e-4b80-8da3-cd3803754ba2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:42Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:42 crc kubenswrapper[4948]: I0312 00:08:42.175578 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:42 crc kubenswrapper[4948]: I0312 00:08:42.175620 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:42 crc kubenswrapper[4948]: I0312 00:08:42.175632 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:42 crc kubenswrapper[4948]: I0312 00:08:42.175650 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:42 crc kubenswrapper[4948]: I0312 00:08:42.175662 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:42Z","lastTransitionTime":"2026-03-12T00:08:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:42 crc kubenswrapper[4948]: E0312 00:08:42.191656 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17825588-29cf-4059-bab7-1b042b9bf1f3\\\",\\\"systemUUID\\\":\\\"b18e4bf2-d37e-4b80-8da3-cd3803754ba2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:42Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:42 crc kubenswrapper[4948]: I0312 00:08:42.195261 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:42 crc kubenswrapper[4948]: I0312 00:08:42.195315 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:42 crc kubenswrapper[4948]: I0312 00:08:42.195330 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:42 crc kubenswrapper[4948]: I0312 00:08:42.195346 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:42 crc kubenswrapper[4948]: I0312 00:08:42.195357 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:42Z","lastTransitionTime":"2026-03-12T00:08:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:42 crc kubenswrapper[4948]: E0312 00:08:42.211855 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17825588-29cf-4059-bab7-1b042b9bf1f3\\\",\\\"systemUUID\\\":\\\"b18e4bf2-d37e-4b80-8da3-cd3803754ba2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:42Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:42 crc kubenswrapper[4948]: I0312 00:08:42.218647 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:42 crc kubenswrapper[4948]: I0312 00:08:42.218698 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:42 crc kubenswrapper[4948]: I0312 00:08:42.218716 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:42 crc kubenswrapper[4948]: I0312 00:08:42.218736 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:42 crc kubenswrapper[4948]: I0312 00:08:42.218753 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:42Z","lastTransitionTime":"2026-03-12T00:08:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:42 crc kubenswrapper[4948]: E0312 00:08:42.248505 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17825588-29cf-4059-bab7-1b042b9bf1f3\\\",\\\"systemUUID\\\":\\\"b18e4bf2-d37e-4b80-8da3-cd3803754ba2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:42Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:42 crc kubenswrapper[4948]: I0312 00:08:42.263332 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:42 crc kubenswrapper[4948]: I0312 00:08:42.263367 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:42 crc kubenswrapper[4948]: I0312 00:08:42.263376 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:42 crc kubenswrapper[4948]: I0312 00:08:42.263390 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:42 crc kubenswrapper[4948]: I0312 00:08:42.263399 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:42Z","lastTransitionTime":"2026-03-12T00:08:42Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:42 crc kubenswrapper[4948]: E0312 00:08:42.276805 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:42Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:42Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:42Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:42Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17825588-29cf-4059-bab7-1b042b9bf1f3\\\",\\\"systemUUID\\\":\\\"b18e4bf2-d37e-4b80-8da3-cd3803754ba2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:42Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:42 crc kubenswrapper[4948]: E0312 00:08:42.277202 4948 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 12 00:08:42 crc kubenswrapper[4948]: I0312 00:08:42.312119 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:08:42 crc kubenswrapper[4948]: E0312 00:08:42.312425 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 00:08:42 crc kubenswrapper[4948]: I0312 00:08:42.312212 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:08:42 crc kubenswrapper[4948]: I0312 00:08:42.312219 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:08:42 crc kubenswrapper[4948]: I0312 00:08:42.312170 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:08:42 crc kubenswrapper[4948]: E0312 00:08:42.312901 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 00:08:42 crc kubenswrapper[4948]: E0312 00:08:42.312748 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mqmtd" podUID="be985ff1-de8a-431f-a230-67894ebe52d5" Mar 12 00:08:42 crc kubenswrapper[4948]: E0312 00:08:42.312991 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 00:08:42 crc kubenswrapper[4948]: I0312 00:08:42.758402 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f7mp5_15117613-9fad-48c7-98c4-a2d84502ded9/ovnkube-controller/2.log" Mar 12 00:08:42 crc kubenswrapper[4948]: I0312 00:08:42.759454 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f7mp5_15117613-9fad-48c7-98c4-a2d84502ded9/ovnkube-controller/1.log" Mar 12 00:08:42 crc kubenswrapper[4948]: I0312 00:08:42.763546 4948 generic.go:334] "Generic (PLEG): container finished" podID="15117613-9fad-48c7-98c4-a2d84502ded9" containerID="a18330e68c4cbdf0380d4f4845bed125a80893c948f89c674632af8eb1ffb5f4" exitCode=1 Mar 12 00:08:42 crc kubenswrapper[4948]: I0312 00:08:42.763601 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" event={"ID":"15117613-9fad-48c7-98c4-a2d84502ded9","Type":"ContainerDied","Data":"a18330e68c4cbdf0380d4f4845bed125a80893c948f89c674632af8eb1ffb5f4"} Mar 12 00:08:42 crc kubenswrapper[4948]: I0312 00:08:42.763650 4948 scope.go:117] "RemoveContainer" containerID="588e8a26bafbb970547438b573ab34f111ef18d8134df2b3fff513c9cca9a516" Mar 12 00:08:42 crc kubenswrapper[4948]: I0312 00:08:42.765108 4948 scope.go:117] "RemoveContainer" containerID="a18330e68c4cbdf0380d4f4845bed125a80893c948f89c674632af8eb1ffb5f4" Mar 12 00:08:42 crc kubenswrapper[4948]: E0312 00:08:42.766451 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-f7mp5_openshift-ovn-kubernetes(15117613-9fad-48c7-98c4-a2d84502ded9)\"" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" podUID="15117613-9fad-48c7-98c4-a2d84502ded9" Mar 12 00:08:42 crc kubenswrapper[4948]: I0312 00:08:42.788157 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5139686878eb07d3abced92d0ccfe05be8907580ed75233b72fc2ad3415a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15cbe3fa75b69a7176803295e16a314a8bc73f5797c874124d54e76d47fa1064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:42Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:42 crc kubenswrapper[4948]: I0312 00:08:42.806492 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:42Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:42 crc kubenswrapper[4948]: I0312 00:08:42.828272 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfp2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"662dcbe7-7cf8-42d4-9a5e-345c32a49972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff3f3c69dcb28c7b0fd85f9e0c33b3339a1f6159dbf10bcf0efbdce4f7728c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a57d476b5b00110168878bd1c19c32ef122525c0259ac0a70de31c8220996d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4a57d476b5b00110168878bd1c19c32ef122525c0259ac0a70de31c8220996d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bb429b94a1286d8aac77cb77112c7163f7aa50a1c9cab997f8d088c100887c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0bb429b94a1286d8aac77cb77112c7163f7aa50a1c9cab997f8d088c100887c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a77485fd9e246f190724b86663e0dbe81146e1fcc51f468dd05f23c1db878fd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a77485fd9e246f190724b86663e0dbe81146e1fcc51f468dd05f23c1db878fd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adaa20d9a4898057802a8c9759ab3cd485963180bfc4f6b926bf6d2c886bcf27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adaa20d9a4898057802a8c9759ab3cd485963180bfc4f6b926bf6d2c886bcf27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://998fbb934f01c9ecd315f78e776199d4fd538106f8373f09b730d2df23ac9e3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://998fbb934f01c9ecd315f78e776199d4fd538106f8373f09b730d2df23ac9e3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46ebe18804ff07eb0e922a6cc1c39ff93875302d231222642b211e3527fe6373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46ebe18804ff07eb0e922a6cc1c39ff93875302d231222642b211e3527fe6373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfp2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:42Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:42 crc kubenswrapper[4948]: I0312 00:08:42.843732 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e483c2b-08f0-4e92-8e4a-b7281f30af3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d6839c99ab958ad1f40636c70e29bc285292fdbee1e910cdf7f8ab43ca7e8db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd9jv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82de8f2b633f1b6c295149d077bfd2b308f5acaa429acf411ed16f456536d0ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd9jv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4xwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:42Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:42 crc kubenswrapper[4948]: I0312 00:08:42.863224 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c835bb7a-4e06-4047-9fe6-e0c605b36bc6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc9760bd6ed0d092081fdd2ad33dda2d67f1305531640a35b9dce84b5d86a813\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f580535b16c31fb0ebadbdd9b7851ba2dce2f7d52bcad8c9cacc197f5e7dfad4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T00:07:08Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0312 00:06:43.492131 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0312 00:06:43.494442 1 observer_polling.go:159] Starting file observer\\\\nI0312 00:06:43.528056 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0312 00:06:43.536640 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0312 00:07:08.239474 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0312 00:07:08.239540 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb45482a404d6696cfbc3919997e53146b7b9dd1536056a55374574ae01ff193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://663bca87e44109c1d2bb59964e567d8b78fe2e9de39bcec0557cb43e39143add\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bf5b5d40a0cc76725ae2cf453089b31e75c304cf6037ef06b8ca8f7cff35a5a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:06:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:42Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:42 crc kubenswrapper[4948]: I0312 00:08:42.883940 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2efff53-8e12-41a1-bcb1-9a1d08c42a4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1cb328c8431b39dac735d49726a2eae40a367849b2c146bb2e49d02958f0e5a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ad13662849131222d5c8ba1041b81fa75bcb1fcbfa2e71c0c751f04b532a5b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8cd4faf47b25395ffd42d4bf5416ea29dcaa3ec380d6d6d6a9f13321d35195a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8028fd7b3e6d26bf30666425c843933b8fb87c3b2a258d510608ec718eb4fe82\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73834c8db23f6536034dce09e38b56665211238280bf9a4e1b13e74d81245a77\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T00:07:50Z\\\",\\\"message\\\":\\\"le observer\\\\nW0312 00:07:49.927227 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0312 00:07:49.927347 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 00:07:49.928000 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2511358960/tls.crt::/tmp/serving-cert-2511358960/tls.key\\\\\\\"\\\\nI0312 00:07:50.321455 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0312 00:07:50.324479 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0312 00:07:50.324511 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0312 00:07:50.324546 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0312 00:07:50.324556 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0312 00:07:50.331591 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0312 00:07:50.331609 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0312 00:07:50.331634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 00:07:50.331642 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 00:07:50.331651 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0312 00:07:50.331658 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0312 00:07:50.331664 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0312 00:07:50.331670 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0312 00:07:50.332917 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:07:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d56e32f9cb96ad103bab80d6a730b70d3387264a25a7a61ef1f0726e7290d6c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e42f2e94463ac8a3d7980a25122062da9ee87fcdc1f10e260ff9974b2e42359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e42f2e94463ac8a3d7980a25122062da9ee87fcdc1f10e260ff9974b2e42359\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:06:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:42Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:42 crc kubenswrapper[4948]: I0312 00:08:42.905638 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5a9ae9441886817563a5fc793bff7c9b8d77d406ceb1d1fb1eebd4307a1c64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:42Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:42 crc kubenswrapper[4948]: I0312 00:08:42.925096 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwpqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b7580c2-5c23-4c67-807a-ea97a3df9398\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d07c603619f858177d4b007e785208853abcfe2dc6c1af435309030fa409010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8lxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwpqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:42Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:42 crc kubenswrapper[4948]: I0312 00:08:42.940433 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-njc9c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0d6440c-4524-42ea-b18a-0ddc4662f672\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f4d0d9599ad62aaba35ba1414a806f3a163e29a3c34fd893ef5b14eae003591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7wn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-njc9c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:42Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:42 crc kubenswrapper[4948]: I0312 00:08:42.955836 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9s5fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f9c7f1-6593-4d92-854c-e6c9964a0435\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b8833cbaf3809463e18bfebc9d5bada5fb26823831cc742a3a27177662f60e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xf4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9s5fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:42Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:42 crc kubenswrapper[4948]: I0312 00:08:42.972090 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xdd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b5233d6-d2b8-42dc-b881-18e439ecc442\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0ea55d993c72b6d50a4829ecfff4adae7d0f625a90e40545c9e63201cdda29c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjpwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb23ff4bc484524e306561c0c574e2b2eee39fc80efcd9f44f1924fe42df1a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjpwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4xdd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:42Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:42 crc kubenswrapper[4948]: I0312 00:08:42.989078 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a124d35d-f697-49ae-8dac-ee32cd6fc5f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49c25293fe6ddb25b23fc250e3a5d94ce3f2c464f450c0da24dcff2febf15124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a57137a3fb48ab82c4324c6121b7dbcde8dd33e02cd7012ff42248cc5f164bec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccb319a44802931d007a0d9265f98bac6f23efee9a7fb82798718a315464e89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c756f1a5b429d17d95df4797725081682b66e9ef3e53a9deb9518e326e40325b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c756f1a5b429d17d95df4797725081682b66e9ef3e53a9deb9518e326e40325b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:06:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:42Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:43 crc kubenswrapper[4948]: I0312 00:08:43.007097 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:43Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:43 crc kubenswrapper[4948]: I0312 00:08:43.025228 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8d58ee45d75284e030700fa526b2c39197a26f6a391a7417f25a026f356b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:43Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:43 crc kubenswrapper[4948]: I0312 00:08:43.041646 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mqmtd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be985ff1-de8a-431f-a230-67894ebe52d5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmsz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmsz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mqmtd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:43Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:43 crc kubenswrapper[4948]: I0312 00:08:43.073868 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15117613-9fad-48c7-98c4-a2d84502ded9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4036d7131db00afd8eab8e093626b688297dc776b5af80c3c38b51da7124085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84af4cdd213d90636a889a8a8a3aaca3dee1e2677368efa0ffc768940c2cc994\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e481ace9e84b6aa66cbdda3ee8425c4120a665915d2c70ae4faf60bba6205e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9eadefd0babc119f776d48c9d2c9e4bced58db685af9a8e97f25e6a9a2e3c89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9d5e77753fda70350353daa0967257dd4929dba4440a95495ab4c2a0659aef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cad943edb9a8d445b33a6e70d85b475c1ccba460a6056bccae94fe60fd6b73f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a18330e68c4cbdf0380d4f4845bed125a80893c948f89c674632af8eb1ffb5f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://588e8a26bafbb970547438b573ab34f111ef18d8134df2b3fff513c9cca9a516\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T00:08:25Z\\\",\\\"message\\\":\\\"umn _uuid == {f6d604c1-9711-4e25-be6c-79ec28bbad1b}] Until: Durable:\\\\u003cnil\\\\u003e Comment:\\\\u003cnil\\\\u003e Lock:\\\\u003cnil\\\\u003e UUID: UUIDName:}]\\\\nI0312 00:08:25.737942 6942 obj_retry.go:551] Creating *factory.egressNode crc took: 2.187304ms\\\\nI0312 00:08:25.737970 6942 factory.go:1336] Added *v1.Node event handler 7\\\\nI0312 00:08:25.738003 6942 factory.go:1336] Added *v1.EgressIP event handler 8\\\\nI0312 00:08:25.738210 6942 factory.go:1336] Added *v1.EgressFirewall event handler 9\\\\nI0312 00:08:25.738274 6942 controller.go:132] Adding controller ef_node_controller event handlers\\\\nI0312 00:08:25.738325 6942 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0312 00:08:25.738333 6942 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0312 00:08:25.738357 6942 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0312 00:08:25.738373 6942 factory.go:656] Stopping watch factory\\\\nI0312 00:08:25.738386 6942 ovnkube.go:599] Stopped ovnkube\\\\nI0312 00:08:25.738410 6942 handler.go:208] Removed *v1.Node event handler 2\\\\nI0312 00:08:25.738414 6942 metrics.go:553] Stopping metrics server at address \\\\\\\"127.0.0.1:29103\\\\\\\"\\\\nI0312 00:08:25.738430 6942 handler.go:208] Removed *v1.Node event handler 7\\\\nI0312 00:08:25.738442 6942 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nF0312 00:08:25.738550 6942 ovnkube.go:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:24Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a18330e68c4cbdf0380d4f4845bed125a80893c948f89c674632af8eb1ffb5f4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T00:08:42Z\\\",\\\"message\\\":\\\"lector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0312 00:08:42.392052 7170 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0312 00:08:42.392236 7170 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0312 00:08:42.392701 7170 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0312 00:08:42.392763 7170 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0312 00:08:42.392775 7170 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0312 00:08:42.392819 7170 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0312 00:08:42.392895 7170 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0312 00:08:42.392966 7170 factory.go:656] Stopping watch factory\\\\nI0312 00:08:42.392993 7170 ovnkube.go:599] Stopped ovnkube\\\\nI0312 00:08:42.393027 7170 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0312 00:08:42.393049 7170 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0312 00:08:42.393062 7170 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0312 00:08:42.393074 7170 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0312 00:08:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:41Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46fa6140f4f4bd32e298fa36f6d7af6890029f9bc3080d2b4d94be8307346de8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7mp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:43Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:43 crc kubenswrapper[4948]: I0312 00:08:43.093363 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:43Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:43 crc kubenswrapper[4948]: I0312 00:08:43.768522 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f7mp5_15117613-9fad-48c7-98c4-a2d84502ded9/ovnkube-controller/2.log" Mar 12 00:08:43 crc kubenswrapper[4948]: I0312 00:08:43.773479 4948 scope.go:117] "RemoveContainer" containerID="a18330e68c4cbdf0380d4f4845bed125a80893c948f89c674632af8eb1ffb5f4" Mar 12 00:08:43 crc kubenswrapper[4948]: E0312 00:08:43.773732 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-f7mp5_openshift-ovn-kubernetes(15117613-9fad-48c7-98c4-a2d84502ded9)\"" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" podUID="15117613-9fad-48c7-98c4-a2d84502ded9" Mar 12 00:08:43 crc kubenswrapper[4948]: I0312 00:08:43.795282 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2efff53-8e12-41a1-bcb1-9a1d08c42a4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1cb328c8431b39dac735d49726a2eae40a367849b2c146bb2e49d02958f0e5a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ad13662849131222d5c8ba1041b81fa75bcb1fcbfa2e71c0c751f04b532a5b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8cd4faf47b25395ffd42d4bf5416ea29dcaa3ec380d6d6d6a9f13321d35195a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8028fd7b3e6d26bf30666425c843933b8fb87c3b2a258d510608ec718eb4fe82\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73834c8db23f6536034dce09e38b56665211238280bf9a4e1b13e74d81245a77\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T00:07:50Z\\\",\\\"message\\\":\\\"le observer\\\\nW0312 00:07:49.927227 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0312 00:07:49.927347 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 00:07:49.928000 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2511358960/tls.crt::/tmp/serving-cert-2511358960/tls.key\\\\\\\"\\\\nI0312 00:07:50.321455 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0312 00:07:50.324479 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0312 00:07:50.324511 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0312 00:07:50.324546 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0312 00:07:50.324556 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0312 00:07:50.331591 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0312 00:07:50.331609 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0312 00:07:50.331634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 00:07:50.331642 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 00:07:50.331651 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0312 00:07:50.331658 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0312 00:07:50.331664 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0312 00:07:50.331670 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0312 00:07:50.332917 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:07:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d56e32f9cb96ad103bab80d6a730b70d3387264a25a7a61ef1f0726e7290d6c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e42f2e94463ac8a3d7980a25122062da9ee87fcdc1f10e260ff9974b2e42359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e42f2e94463ac8a3d7980a25122062da9ee87fcdc1f10e260ff9974b2e42359\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:06:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:43Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:43 crc kubenswrapper[4948]: I0312 00:08:43.818194 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5a9ae9441886817563a5fc793bff7c9b8d77d406ceb1d1fb1eebd4307a1c64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:43Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:43 crc kubenswrapper[4948]: I0312 00:08:43.835392 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwpqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b7580c2-5c23-4c67-807a-ea97a3df9398\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d07c603619f858177d4b007e785208853abcfe2dc6c1af435309030fa409010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8lxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwpqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:43Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:43 crc kubenswrapper[4948]: I0312 00:08:43.849451 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-njc9c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0d6440c-4524-42ea-b18a-0ddc4662f672\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f4d0d9599ad62aaba35ba1414a806f3a163e29a3c34fd893ef5b14eae003591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7wn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-njc9c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:43Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:43 crc kubenswrapper[4948]: I0312 00:08:43.865431 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a124d35d-f697-49ae-8dac-ee32cd6fc5f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49c25293fe6ddb25b23fc250e3a5d94ce3f2c464f450c0da24dcff2febf15124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a57137a3fb48ab82c4324c6121b7dbcde8dd33e02cd7012ff42248cc5f164bec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccb319a44802931d007a0d9265f98bac6f23efee9a7fb82798718a315464e89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c756f1a5b429d17d95df4797725081682b66e9ef3e53a9deb9518e326e40325b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c756f1a5b429d17d95df4797725081682b66e9ef3e53a9deb9518e326e40325b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:06:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:43Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:43 crc kubenswrapper[4948]: I0312 00:08:43.879176 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9s5fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f9c7f1-6593-4d92-854c-e6c9964a0435\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b8833cbaf3809463e18bfebc9d5bada5fb26823831cc742a3a27177662f60e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xf4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9s5fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:43Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:43 crc kubenswrapper[4948]: I0312 00:08:43.896562 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xdd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b5233d6-d2b8-42dc-b881-18e439ecc442\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0ea55d993c72b6d50a4829ecfff4adae7d0f625a90e40545c9e63201cdda29c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjpwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb23ff4bc484524e306561c0c574e2b2eee39fc80efcd9f44f1924fe42df1a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjpwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4xdd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:43Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:43 crc kubenswrapper[4948]: I0312 00:08:43.916975 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:43Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:43 crc kubenswrapper[4948]: I0312 00:08:43.942185 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:43Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:43 crc kubenswrapper[4948]: I0312 00:08:43.962000 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8d58ee45d75284e030700fa526b2c39197a26f6a391a7417f25a026f356b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:43Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:43 crc kubenswrapper[4948]: I0312 00:08:43.979536 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mqmtd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be985ff1-de8a-431f-a230-67894ebe52d5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmsz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmsz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mqmtd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:43Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:44 crc kubenswrapper[4948]: I0312 00:08:44.001098 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15117613-9fad-48c7-98c4-a2d84502ded9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4036d7131db00afd8eab8e093626b688297dc776b5af80c3c38b51da7124085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84af4cdd213d90636a889a8a8a3aaca3dee1e2677368efa0ffc768940c2cc994\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e481ace9e84b6aa66cbdda3ee8425c4120a665915d2c70ae4faf60bba6205e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9eadefd0babc119f776d48c9d2c9e4bced58db685af9a8e97f25e6a9a2e3c89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9d5e77753fda70350353daa0967257dd4929dba4440a95495ab4c2a0659aef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cad943edb9a8d445b33a6e70d85b475c1ccba460a6056bccae94fe60fd6b73f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a18330e68c4cbdf0380d4f4845bed125a80893c948f89c674632af8eb1ffb5f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a18330e68c4cbdf0380d4f4845bed125a80893c948f89c674632af8eb1ffb5f4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T00:08:42Z\\\",\\\"message\\\":\\\"lector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0312 00:08:42.392052 7170 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0312 00:08:42.392236 7170 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0312 00:08:42.392701 7170 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0312 00:08:42.392763 7170 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0312 00:08:42.392775 7170 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0312 00:08:42.392819 7170 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0312 00:08:42.392895 7170 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0312 00:08:42.392966 7170 factory.go:656] Stopping watch factory\\\\nI0312 00:08:42.392993 7170 ovnkube.go:599] Stopped ovnkube\\\\nI0312 00:08:42.393027 7170 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0312 00:08:42.393049 7170 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0312 00:08:42.393062 7170 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0312 00:08:42.393074 7170 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0312 00:08:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-f7mp5_openshift-ovn-kubernetes(15117613-9fad-48c7-98c4-a2d84502ded9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46fa6140f4f4bd32e298fa36f6d7af6890029f9bc3080d2b4d94be8307346de8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7mp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:43Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:44 crc kubenswrapper[4948]: I0312 00:08:44.018690 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c835bb7a-4e06-4047-9fe6-e0c605b36bc6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc9760bd6ed0d092081fdd2ad33dda2d67f1305531640a35b9dce84b5d86a813\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f580535b16c31fb0ebadbdd9b7851ba2dce2f7d52bcad8c9cacc197f5e7dfad4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T00:07:08Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0312 00:06:43.492131 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0312 00:06:43.494442 1 observer_polling.go:159] Starting file observer\\\\nI0312 00:06:43.528056 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0312 00:06:43.536640 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0312 00:07:08.239474 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0312 00:07:08.239540 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb45482a404d6696cfbc3919997e53146b7b9dd1536056a55374574ae01ff193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://663bca87e44109c1d2bb59964e567d8b78fe2e9de39bcec0557cb43e39143add\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bf5b5d40a0cc76725ae2cf453089b31e75c304cf6037ef06b8ca8f7cff35a5a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:06:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:44Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:44 crc kubenswrapper[4948]: I0312 00:08:44.036605 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5139686878eb07d3abced92d0ccfe05be8907580ed75233b72fc2ad3415a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15cbe3fa75b69a7176803295e16a314a8bc73f5797c874124d54e76d47fa1064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:44Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:44 crc kubenswrapper[4948]: I0312 00:08:44.053981 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:44Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:44 crc kubenswrapper[4948]: I0312 00:08:44.072256 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfp2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"662dcbe7-7cf8-42d4-9a5e-345c32a49972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff3f3c69dcb28c7b0fd85f9e0c33b3339a1f6159dbf10bcf0efbdce4f7728c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a57d476b5b00110168878bd1c19c32ef122525c0259ac0a70de31c8220996d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4a57d476b5b00110168878bd1c19c32ef122525c0259ac0a70de31c8220996d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bb429b94a1286d8aac77cb77112c7163f7aa50a1c9cab997f8d088c100887c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0bb429b94a1286d8aac77cb77112c7163f7aa50a1c9cab997f8d088c100887c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a77485fd9e246f190724b86663e0dbe81146e1fcc51f468dd05f23c1db878fd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a77485fd9e246f190724b86663e0dbe81146e1fcc51f468dd05f23c1db878fd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adaa20d9a4898057802a8c9759ab3cd485963180bfc4f6b926bf6d2c886bcf27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adaa20d9a4898057802a8c9759ab3cd485963180bfc4f6b926bf6d2c886bcf27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://998fbb934f01c9ecd315f78e776199d4fd538106f8373f09b730d2df23ac9e3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://998fbb934f01c9ecd315f78e776199d4fd538106f8373f09b730d2df23ac9e3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46ebe18804ff07eb0e922a6cc1c39ff93875302d231222642b211e3527fe6373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46ebe18804ff07eb0e922a6cc1c39ff93875302d231222642b211e3527fe6373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfp2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:44Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:44 crc kubenswrapper[4948]: I0312 00:08:44.086564 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e483c2b-08f0-4e92-8e4a-b7281f30af3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d6839c99ab958ad1f40636c70e29bc285292fdbee1e910cdf7f8ab43ca7e8db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd9jv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82de8f2b633f1b6c295149d077bfd2b308f5acaa429acf411ed16f456536d0ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd9jv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4xwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:44Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:44 crc kubenswrapper[4948]: I0312 00:08:44.312239 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:08:44 crc kubenswrapper[4948]: I0312 00:08:44.312277 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:08:44 crc kubenswrapper[4948]: I0312 00:08:44.312431 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:08:44 crc kubenswrapper[4948]: E0312 00:08:44.312606 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 00:08:44 crc kubenswrapper[4948]: E0312 00:08:44.312739 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 00:08:44 crc kubenswrapper[4948]: E0312 00:08:44.312820 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 00:08:44 crc kubenswrapper[4948]: I0312 00:08:44.313079 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:08:44 crc kubenswrapper[4948]: E0312 00:08:44.313471 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mqmtd" podUID="be985ff1-de8a-431f-a230-67894ebe52d5" Mar 12 00:08:46 crc kubenswrapper[4948]: I0312 00:08:46.127038 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 00:08:46 crc kubenswrapper[4948]: E0312 00:08:46.127356 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 00:09:18.12727175 +0000 UTC m=+157.582875528 (durationBeforeRetry 32s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:08:46 crc kubenswrapper[4948]: I0312 00:08:46.127441 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:08:46 crc kubenswrapper[4948]: I0312 00:08:46.127502 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:08:46 crc kubenswrapper[4948]: I0312 00:08:46.127558 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:08:46 crc kubenswrapper[4948]: I0312 00:08:46.127598 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:08:46 crc kubenswrapper[4948]: E0312 00:08:46.127660 4948 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 12 00:08:46 crc kubenswrapper[4948]: E0312 00:08:46.127740 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 12 00:08:46 crc kubenswrapper[4948]: E0312 00:08:46.127748 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 12 00:08:46 crc kubenswrapper[4948]: E0312 00:08:46.127782 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 12 00:08:46 crc kubenswrapper[4948]: E0312 00:08:46.127768 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 12 00:08:46 crc kubenswrapper[4948]: E0312 00:08:46.127823 4948 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 00:08:46 crc kubenswrapper[4948]: E0312 00:08:46.127800 4948 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 00:08:46 crc kubenswrapper[4948]: E0312 00:08:46.127751 4948 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 12 00:08:46 crc kubenswrapper[4948]: E0312 00:08:46.127742 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-12 00:09:18.127717991 +0000 UTC m=+157.583321759 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 12 00:08:46 crc kubenswrapper[4948]: E0312 00:08:46.127994 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-12 00:09:18.127971893 +0000 UTC m=+157.583575661 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 00:08:46 crc kubenswrapper[4948]: E0312 00:08:46.128015 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-12 00:09:18.128004114 +0000 UTC m=+157.583607882 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 00:08:46 crc kubenswrapper[4948]: E0312 00:08:46.128035 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-12 00:09:18.128024595 +0000 UTC m=+157.583628373 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 12 00:08:46 crc kubenswrapper[4948]: I0312 00:08:46.228983 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/be985ff1-de8a-431f-a230-67894ebe52d5-metrics-certs\") pod \"network-metrics-daemon-mqmtd\" (UID: \"be985ff1-de8a-431f-a230-67894ebe52d5\") " pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:08:46 crc kubenswrapper[4948]: E0312 00:08:46.229201 4948 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 12 00:08:46 crc kubenswrapper[4948]: E0312 00:08:46.229305 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/be985ff1-de8a-431f-a230-67894ebe52d5-metrics-certs podName:be985ff1-de8a-431f-a230-67894ebe52d5 nodeName:}" failed. No retries permitted until 2026-03-12 00:09:18.229275011 +0000 UTC m=+157.684878789 (durationBeforeRetry 32s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/be985ff1-de8a-431f-a230-67894ebe52d5-metrics-certs") pod "network-metrics-daemon-mqmtd" (UID: "be985ff1-de8a-431f-a230-67894ebe52d5") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 12 00:08:46 crc kubenswrapper[4948]: I0312 00:08:46.312128 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:08:46 crc kubenswrapper[4948]: I0312 00:08:46.312153 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:08:46 crc kubenswrapper[4948]: E0312 00:08:46.312358 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 00:08:46 crc kubenswrapper[4948]: I0312 00:08:46.312424 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:08:46 crc kubenswrapper[4948]: I0312 00:08:46.312423 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:08:46 crc kubenswrapper[4948]: E0312 00:08:46.312527 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mqmtd" podUID="be985ff1-de8a-431f-a230-67894ebe52d5" Mar 12 00:08:46 crc kubenswrapper[4948]: E0312 00:08:46.312683 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 00:08:46 crc kubenswrapper[4948]: E0312 00:08:46.312772 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 00:08:46 crc kubenswrapper[4948]: E0312 00:08:46.436731 4948 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 12 00:08:48 crc kubenswrapper[4948]: I0312 00:08:48.312238 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:08:48 crc kubenswrapper[4948]: I0312 00:08:48.312288 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:08:48 crc kubenswrapper[4948]: E0312 00:08:48.312789 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 00:08:48 crc kubenswrapper[4948]: I0312 00:08:48.312460 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:08:48 crc kubenswrapper[4948]: I0312 00:08:48.312294 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:08:48 crc kubenswrapper[4948]: E0312 00:08:48.312988 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 00:08:48 crc kubenswrapper[4948]: E0312 00:08:48.313192 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 00:08:48 crc kubenswrapper[4948]: E0312 00:08:48.313441 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mqmtd" podUID="be985ff1-de8a-431f-a230-67894ebe52d5" Mar 12 00:08:50 crc kubenswrapper[4948]: I0312 00:08:50.312403 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:08:50 crc kubenswrapper[4948]: I0312 00:08:50.312448 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:08:50 crc kubenswrapper[4948]: I0312 00:08:50.312490 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:08:50 crc kubenswrapper[4948]: E0312 00:08:50.312619 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 00:08:50 crc kubenswrapper[4948]: I0312 00:08:50.312468 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:08:50 crc kubenswrapper[4948]: E0312 00:08:50.312877 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mqmtd" podUID="be985ff1-de8a-431f-a230-67894ebe52d5" Mar 12 00:08:50 crc kubenswrapper[4948]: E0312 00:08:50.313003 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 00:08:50 crc kubenswrapper[4948]: E0312 00:08:50.313170 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 00:08:51 crc kubenswrapper[4948]: I0312 00:08:51.334261 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a124d35d-f697-49ae-8dac-ee32cd6fc5f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49c25293fe6ddb25b23fc250e3a5d94ce3f2c464f450c0da24dcff2febf15124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a57137a3fb48ab82c4324c6121b7dbcde8dd33e02cd7012ff42248cc5f164bec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccb319a44802931d007a0d9265f98bac6f23efee9a7fb82798718a315464e89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c756f1a5b429d17d95df4797725081682b66e9ef3e53a9deb9518e326e40325b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c756f1a5b429d17d95df4797725081682b66e9ef3e53a9deb9518e326e40325b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:06:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:51Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:51 crc kubenswrapper[4948]: I0312 00:08:51.350427 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9s5fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f9c7f1-6593-4d92-854c-e6c9964a0435\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b8833cbaf3809463e18bfebc9d5bada5fb26823831cc742a3a27177662f60e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xf4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9s5fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:51Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:51 crc kubenswrapper[4948]: I0312 00:08:51.368732 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xdd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b5233d6-d2b8-42dc-b881-18e439ecc442\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0ea55d993c72b6d50a4829ecfff4adae7d0f625a90e40545c9e63201cdda29c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjpwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb23ff4bc484524e306561c0c574e2b2eee39fc80efcd9f44f1924fe42df1a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjpwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4xdd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:51Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:51 crc kubenswrapper[4948]: I0312 00:08:51.401886 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15117613-9fad-48c7-98c4-a2d84502ded9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4036d7131db00afd8eab8e093626b688297dc776b5af80c3c38b51da7124085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84af4cdd213d90636a889a8a8a3aaca3dee1e2677368efa0ffc768940c2cc994\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e481ace9e84b6aa66cbdda3ee8425c4120a665915d2c70ae4faf60bba6205e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9eadefd0babc119f776d48c9d2c9e4bced58db685af9a8e97f25e6a9a2e3c89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9d5e77753fda70350353daa0967257dd4929dba4440a95495ab4c2a0659aef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cad943edb9a8d445b33a6e70d85b475c1ccba460a6056bccae94fe60fd6b73f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a18330e68c4cbdf0380d4f4845bed125a80893c948f89c674632af8eb1ffb5f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a18330e68c4cbdf0380d4f4845bed125a80893c948f89c674632af8eb1ffb5f4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T00:08:42Z\\\",\\\"message\\\":\\\"lector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0312 00:08:42.392052 7170 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0312 00:08:42.392236 7170 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0312 00:08:42.392701 7170 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0312 00:08:42.392763 7170 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0312 00:08:42.392775 7170 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0312 00:08:42.392819 7170 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0312 00:08:42.392895 7170 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0312 00:08:42.392966 7170 factory.go:656] Stopping watch factory\\\\nI0312 00:08:42.392993 7170 ovnkube.go:599] Stopped ovnkube\\\\nI0312 00:08:42.393027 7170 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0312 00:08:42.393049 7170 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0312 00:08:42.393062 7170 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0312 00:08:42.393074 7170 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0312 00:08:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-f7mp5_openshift-ovn-kubernetes(15117613-9fad-48c7-98c4-a2d84502ded9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46fa6140f4f4bd32e298fa36f6d7af6890029f9bc3080d2b4d94be8307346de8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7mp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:51Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:51 crc kubenswrapper[4948]: I0312 00:08:51.425014 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:51Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:51 crc kubenswrapper[4948]: E0312 00:08:51.437694 4948 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 12 00:08:51 crc kubenswrapper[4948]: I0312 00:08:51.449957 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:51Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:51 crc kubenswrapper[4948]: I0312 00:08:51.469894 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8d58ee45d75284e030700fa526b2c39197a26f6a391a7417f25a026f356b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:51Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:51 crc kubenswrapper[4948]: I0312 00:08:51.485268 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mqmtd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be985ff1-de8a-431f-a230-67894ebe52d5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmsz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmsz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mqmtd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:51Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:51 crc kubenswrapper[4948]: I0312 00:08:51.510169 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfp2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"662dcbe7-7cf8-42d4-9a5e-345c32a49972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff3f3c69dcb28c7b0fd85f9e0c33b3339a1f6159dbf10bcf0efbdce4f7728c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a57d476b5b00110168878bd1c19c32ef122525c0259ac0a70de31c8220996d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4a57d476b5b00110168878bd1c19c32ef122525c0259ac0a70de31c8220996d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bb429b94a1286d8aac77cb77112c7163f7aa50a1c9cab997f8d088c100887c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0bb429b94a1286d8aac77cb77112c7163f7aa50a1c9cab997f8d088c100887c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a77485fd9e246f190724b86663e0dbe81146e1fcc51f468dd05f23c1db878fd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a77485fd9e246f190724b86663e0dbe81146e1fcc51f468dd05f23c1db878fd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adaa20d9a4898057802a8c9759ab3cd485963180bfc4f6b926bf6d2c886bcf27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adaa20d9a4898057802a8c9759ab3cd485963180bfc4f6b926bf6d2c886bcf27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://998fbb934f01c9ecd315f78e776199d4fd538106f8373f09b730d2df23ac9e3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://998fbb934f01c9ecd315f78e776199d4fd538106f8373f09b730d2df23ac9e3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46ebe18804ff07eb0e922a6cc1c39ff93875302d231222642b211e3527fe6373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46ebe18804ff07eb0e922a6cc1c39ff93875302d231222642b211e3527fe6373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfp2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:51Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:51 crc kubenswrapper[4948]: I0312 00:08:51.528553 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e483c2b-08f0-4e92-8e4a-b7281f30af3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d6839c99ab958ad1f40636c70e29bc285292fdbee1e910cdf7f8ab43ca7e8db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd9jv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82de8f2b633f1b6c295149d077bfd2b308f5acaa429acf411ed16f456536d0ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd9jv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4xwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:51Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:51 crc kubenswrapper[4948]: I0312 00:08:51.548833 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c835bb7a-4e06-4047-9fe6-e0c605b36bc6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc9760bd6ed0d092081fdd2ad33dda2d67f1305531640a35b9dce84b5d86a813\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f580535b16c31fb0ebadbdd9b7851ba2dce2f7d52bcad8c9cacc197f5e7dfad4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T00:07:08Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0312 00:06:43.492131 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0312 00:06:43.494442 1 observer_polling.go:159] Starting file observer\\\\nI0312 00:06:43.528056 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0312 00:06:43.536640 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0312 00:07:08.239474 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0312 00:07:08.239540 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb45482a404d6696cfbc3919997e53146b7b9dd1536056a55374574ae01ff193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://663bca87e44109c1d2bb59964e567d8b78fe2e9de39bcec0557cb43e39143add\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bf5b5d40a0cc76725ae2cf453089b31e75c304cf6037ef06b8ca8f7cff35a5a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:06:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:51Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:51 crc kubenswrapper[4948]: I0312 00:08:51.569079 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5139686878eb07d3abced92d0ccfe05be8907580ed75233b72fc2ad3415a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15cbe3fa75b69a7176803295e16a314a8bc73f5797c874124d54e76d47fa1064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:51Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:51 crc kubenswrapper[4948]: I0312 00:08:51.588886 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:51Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:51 crc kubenswrapper[4948]: I0312 00:08:51.604978 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-njc9c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0d6440c-4524-42ea-b18a-0ddc4662f672\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f4d0d9599ad62aaba35ba1414a806f3a163e29a3c34fd893ef5b14eae003591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7wn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-njc9c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:51Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:51 crc kubenswrapper[4948]: I0312 00:08:51.628929 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2efff53-8e12-41a1-bcb1-9a1d08c42a4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-apiserver-check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1cb328c8431b39dac735d49726a2eae40a367849b2c146bb2e49d02958f0e5a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ad13662849131222d5c8ba1041b81fa75bcb1fcbfa2e71c0c751f04b532a5b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8cd4faf47b25395ffd42d4bf5416ea29dcaa3ec380d6d6d6a9f13321d35195a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8028fd7b3e6d26bf30666425c843933b8fb87c3b2a258d510608ec718eb4fe82\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73834c8db23f6536034dce09e38b56665211238280bf9a4e1b13e74d81245a77\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T00:07:50Z\\\",\\\"message\\\":\\\"le observer\\\\nW0312 00:07:49.927227 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0312 00:07:49.927347 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 00:07:49.928000 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2511358960/tls.crt::/tmp/serving-cert-2511358960/tls.key\\\\\\\"\\\\nI0312 00:07:50.321455 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0312 00:07:50.324479 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0312 00:07:50.324511 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0312 00:07:50.324546 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0312 00:07:50.324556 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0312 00:07:50.331591 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0312 00:07:50.331609 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0312 00:07:50.331634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 00:07:50.331642 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 00:07:50.331651 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0312 00:07:50.331658 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0312 00:07:50.331664 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0312 00:07:50.331670 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0312 00:07:50.332917 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:07:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d56e32f9cb96ad103bab80d6a730b70d3387264a25a7a61ef1f0726e7290d6c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e42f2e94463ac8a3d7980a25122062da9ee87fcdc1f10e260ff9974b2e42359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e42f2e94463ac8a3d7980a25122062da9ee87fcdc1f10e260ff9974b2e42359\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:06:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:51Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:51 crc kubenswrapper[4948]: I0312 00:08:51.649823 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5a9ae9441886817563a5fc793bff7c9b8d77d406ceb1d1fb1eebd4307a1c64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:51Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:51 crc kubenswrapper[4948]: I0312 00:08:51.674262 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwpqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b7580c2-5c23-4c67-807a-ea97a3df9398\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d07c603619f858177d4b007e785208853abcfe2dc6c1af435309030fa409010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8lxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwpqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:51Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:52 crc kubenswrapper[4948]: I0312 00:08:52.312518 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:08:52 crc kubenswrapper[4948]: I0312 00:08:52.312574 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:08:52 crc kubenswrapper[4948]: I0312 00:08:52.312644 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:08:52 crc kubenswrapper[4948]: I0312 00:08:52.312708 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:08:52 crc kubenswrapper[4948]: E0312 00:08:52.312917 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 00:08:52 crc kubenswrapper[4948]: E0312 00:08:52.313089 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 00:08:52 crc kubenswrapper[4948]: E0312 00:08:52.313200 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mqmtd" podUID="be985ff1-de8a-431f-a230-67894ebe52d5" Mar 12 00:08:52 crc kubenswrapper[4948]: E0312 00:08:52.313279 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 00:08:52 crc kubenswrapper[4948]: I0312 00:08:52.664428 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:52 crc kubenswrapper[4948]: I0312 00:08:52.664496 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:52 crc kubenswrapper[4948]: I0312 00:08:52.664514 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:52 crc kubenswrapper[4948]: I0312 00:08:52.664538 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:52 crc kubenswrapper[4948]: I0312 00:08:52.664555 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:52Z","lastTransitionTime":"2026-03-12T00:08:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:52 crc kubenswrapper[4948]: E0312 00:08:52.686677 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17825588-29cf-4059-bab7-1b042b9bf1f3\\\",\\\"systemUUID\\\":\\\"b18e4bf2-d37e-4b80-8da3-cd3803754ba2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:52Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:52 crc kubenswrapper[4948]: I0312 00:08:52.692182 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:52 crc kubenswrapper[4948]: I0312 00:08:52.692527 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:52 crc kubenswrapper[4948]: I0312 00:08:52.692603 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:52 crc kubenswrapper[4948]: I0312 00:08:52.692631 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:52 crc kubenswrapper[4948]: I0312 00:08:52.692687 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:52Z","lastTransitionTime":"2026-03-12T00:08:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:52 crc kubenswrapper[4948]: E0312 00:08:52.712926 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17825588-29cf-4059-bab7-1b042b9bf1f3\\\",\\\"systemUUID\\\":\\\"b18e4bf2-d37e-4b80-8da3-cd3803754ba2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:52Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:52 crc kubenswrapper[4948]: I0312 00:08:52.718619 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:52 crc kubenswrapper[4948]: I0312 00:08:52.718688 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:52 crc kubenswrapper[4948]: I0312 00:08:52.718711 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:52 crc kubenswrapper[4948]: I0312 00:08:52.718741 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:52 crc kubenswrapper[4948]: I0312 00:08:52.718762 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:52Z","lastTransitionTime":"2026-03-12T00:08:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:52 crc kubenswrapper[4948]: E0312 00:08:52.738065 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17825588-29cf-4059-bab7-1b042b9bf1f3\\\",\\\"systemUUID\\\":\\\"b18e4bf2-d37e-4b80-8da3-cd3803754ba2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:52Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:52 crc kubenswrapper[4948]: I0312 00:08:52.742910 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:52 crc kubenswrapper[4948]: I0312 00:08:52.742972 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:52 crc kubenswrapper[4948]: I0312 00:08:52.742989 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:52 crc kubenswrapper[4948]: I0312 00:08:52.743015 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:52 crc kubenswrapper[4948]: I0312 00:08:52.743035 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:52Z","lastTransitionTime":"2026-03-12T00:08:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:52 crc kubenswrapper[4948]: E0312 00:08:52.762221 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17825588-29cf-4059-bab7-1b042b9bf1f3\\\",\\\"systemUUID\\\":\\\"b18e4bf2-d37e-4b80-8da3-cd3803754ba2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:52Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:52 crc kubenswrapper[4948]: I0312 00:08:52.766659 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:08:52 crc kubenswrapper[4948]: I0312 00:08:52.766725 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:08:52 crc kubenswrapper[4948]: I0312 00:08:52.766743 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:08:52 crc kubenswrapper[4948]: I0312 00:08:52.766768 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:08:52 crc kubenswrapper[4948]: I0312 00:08:52.766787 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:08:52Z","lastTransitionTime":"2026-03-12T00:08:52Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:08:52 crc kubenswrapper[4948]: E0312 00:08:52.785833 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:52Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:52Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:08:52Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:52Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17825588-29cf-4059-bab7-1b042b9bf1f3\\\",\\\"systemUUID\\\":\\\"b18e4bf2-d37e-4b80-8da3-cd3803754ba2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:52Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:52 crc kubenswrapper[4948]: E0312 00:08:52.786052 4948 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 12 00:08:53 crc kubenswrapper[4948]: I0312 00:08:53.981907 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 00:08:54 crc kubenswrapper[4948]: I0312 00:08:54.001816 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5139686878eb07d3abced92d0ccfe05be8907580ed75233b72fc2ad3415a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15cbe3fa75b69a7176803295e16a314a8bc73f5797c874124d54e76d47fa1064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:53Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:54 crc kubenswrapper[4948]: I0312 00:08:54.024059 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:54Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:54 crc kubenswrapper[4948]: I0312 00:08:54.051118 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfp2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"662dcbe7-7cf8-42d4-9a5e-345c32a49972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff3f3c69dcb28c7b0fd85f9e0c33b3339a1f6159dbf10bcf0efbdce4f7728c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a57d476b5b00110168878bd1c19c32ef122525c0259ac0a70de31c8220996d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4a57d476b5b00110168878bd1c19c32ef122525c0259ac0a70de31c8220996d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bb429b94a1286d8aac77cb77112c7163f7aa50a1c9cab997f8d088c100887c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0bb429b94a1286d8aac77cb77112c7163f7aa50a1c9cab997f8d088c100887c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a77485fd9e246f190724b86663e0dbe81146e1fcc51f468dd05f23c1db878fd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a77485fd9e246f190724b86663e0dbe81146e1fcc51f468dd05f23c1db878fd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adaa20d9a4898057802a8c9759ab3cd485963180bfc4f6b926bf6d2c886bcf27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adaa20d9a4898057802a8c9759ab3cd485963180bfc4f6b926bf6d2c886bcf27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://998fbb934f01c9ecd315f78e776199d4fd538106f8373f09b730d2df23ac9e3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://998fbb934f01c9ecd315f78e776199d4fd538106f8373f09b730d2df23ac9e3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46ebe18804ff07eb0e922a6cc1c39ff93875302d231222642b211e3527fe6373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46ebe18804ff07eb0e922a6cc1c39ff93875302d231222642b211e3527fe6373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfp2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:54Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:54 crc kubenswrapper[4948]: I0312 00:08:54.070028 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e483c2b-08f0-4e92-8e4a-b7281f30af3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d6839c99ab958ad1f40636c70e29bc285292fdbee1e910cdf7f8ab43ca7e8db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd9jv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82de8f2b633f1b6c295149d077bfd2b308f5acaa429acf411ed16f456536d0ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd9jv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4xwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:54Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:54 crc kubenswrapper[4948]: I0312 00:08:54.090219 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c835bb7a-4e06-4047-9fe6-e0c605b36bc6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc9760bd6ed0d092081fdd2ad33dda2d67f1305531640a35b9dce84b5d86a813\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f580535b16c31fb0ebadbdd9b7851ba2dce2f7d52bcad8c9cacc197f5e7dfad4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T00:07:08Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0312 00:06:43.492131 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0312 00:06:43.494442 1 observer_polling.go:159] Starting file observer\\\\nI0312 00:06:43.528056 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0312 00:06:43.536640 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0312 00:07:08.239474 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0312 00:07:08.239540 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb45482a404d6696cfbc3919997e53146b7b9dd1536056a55374574ae01ff193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://663bca87e44109c1d2bb59964e567d8b78fe2e9de39bcec0557cb43e39143add\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bf5b5d40a0cc76725ae2cf453089b31e75c304cf6037ef06b8ca8f7cff35a5a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:06:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:54Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:54 crc kubenswrapper[4948]: I0312 00:08:54.111240 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2efff53-8e12-41a1-bcb1-9a1d08c42a4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1cb328c8431b39dac735d49726a2eae40a367849b2c146bb2e49d02958f0e5a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ad13662849131222d5c8ba1041b81fa75bcb1fcbfa2e71c0c751f04b532a5b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8cd4faf47b25395ffd42d4bf5416ea29dcaa3ec380d6d6d6a9f13321d35195a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8028fd7b3e6d26bf30666425c843933b8fb87c3b2a258d510608ec718eb4fe82\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73834c8db23f6536034dce09e38b56665211238280bf9a4e1b13e74d81245a77\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T00:07:50Z\\\",\\\"message\\\":\\\"le observer\\\\nW0312 00:07:49.927227 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0312 00:07:49.927347 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 00:07:49.928000 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2511358960/tls.crt::/tmp/serving-cert-2511358960/tls.key\\\\\\\"\\\\nI0312 00:07:50.321455 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0312 00:07:50.324479 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0312 00:07:50.324511 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0312 00:07:50.324546 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0312 00:07:50.324556 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0312 00:07:50.331591 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0312 00:07:50.331609 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0312 00:07:50.331634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 00:07:50.331642 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 00:07:50.331651 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0312 00:07:50.331658 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0312 00:07:50.331664 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0312 00:07:50.331670 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0312 00:07:50.332917 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:07:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d56e32f9cb96ad103bab80d6a730b70d3387264a25a7a61ef1f0726e7290d6c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e42f2e94463ac8a3d7980a25122062da9ee87fcdc1f10e260ff9974b2e42359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e42f2e94463ac8a3d7980a25122062da9ee87fcdc1f10e260ff9974b2e42359\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:06:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:54Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:54 crc kubenswrapper[4948]: I0312 00:08:54.134505 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5a9ae9441886817563a5fc793bff7c9b8d77d406ceb1d1fb1eebd4307a1c64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:54Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:54 crc kubenswrapper[4948]: I0312 00:08:54.160702 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwpqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b7580c2-5c23-4c67-807a-ea97a3df9398\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d07c603619f858177d4b007e785208853abcfe2dc6c1af435309030fa409010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8lxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwpqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:54Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:54 crc kubenswrapper[4948]: I0312 00:08:54.179119 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-njc9c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0d6440c-4524-42ea-b18a-0ddc4662f672\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f4d0d9599ad62aaba35ba1414a806f3a163e29a3c34fd893ef5b14eae003591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7wn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-njc9c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:54Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:54 crc kubenswrapper[4948]: I0312 00:08:54.200172 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9s5fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f9c7f1-6593-4d92-854c-e6c9964a0435\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b8833cbaf3809463e18bfebc9d5bada5fb26823831cc742a3a27177662f60e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xf4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9s5fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:54Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:54 crc kubenswrapper[4948]: I0312 00:08:54.219351 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xdd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b5233d6-d2b8-42dc-b881-18e439ecc442\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0ea55d993c72b6d50a4829ecfff4adae7d0f625a90e40545c9e63201cdda29c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjpwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb23ff4bc484524e306561c0c574e2b2eee39fc80efcd9f44f1924fe42df1a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjpwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4xdd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:54Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:54 crc kubenswrapper[4948]: I0312 00:08:54.238034 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a124d35d-f697-49ae-8dac-ee32cd6fc5f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49c25293fe6ddb25b23fc250e3a5d94ce3f2c464f450c0da24dcff2febf15124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a57137a3fb48ab82c4324c6121b7dbcde8dd33e02cd7012ff42248cc5f164bec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccb319a44802931d007a0d9265f98bac6f23efee9a7fb82798718a315464e89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c756f1a5b429d17d95df4797725081682b66e9ef3e53a9deb9518e326e40325b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c756f1a5b429d17d95df4797725081682b66e9ef3e53a9deb9518e326e40325b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:06:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:54Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:54 crc kubenswrapper[4948]: I0312 00:08:54.260787 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:54Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:54 crc kubenswrapper[4948]: I0312 00:08:54.278183 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8d58ee45d75284e030700fa526b2c39197a26f6a391a7417f25a026f356b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:54Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:54 crc kubenswrapper[4948]: I0312 00:08:54.293606 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mqmtd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be985ff1-de8a-431f-a230-67894ebe52d5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmsz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmsz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mqmtd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:54Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:54 crc kubenswrapper[4948]: I0312 00:08:54.311510 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:08:54 crc kubenswrapper[4948]: I0312 00:08:54.311547 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:08:54 crc kubenswrapper[4948]: I0312 00:08:54.311571 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:08:54 crc kubenswrapper[4948]: E0312 00:08:54.311655 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mqmtd" podUID="be985ff1-de8a-431f-a230-67894ebe52d5" Mar 12 00:08:54 crc kubenswrapper[4948]: I0312 00:08:54.311670 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:08:54 crc kubenswrapper[4948]: E0312 00:08:54.311790 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 00:08:54 crc kubenswrapper[4948]: E0312 00:08:54.311883 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 00:08:54 crc kubenswrapper[4948]: E0312 00:08:54.312667 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 00:08:54 crc kubenswrapper[4948]: I0312 00:08:54.312995 4948 scope.go:117] "RemoveContainer" containerID="a18330e68c4cbdf0380d4f4845bed125a80893c948f89c674632af8eb1ffb5f4" Mar 12 00:08:54 crc kubenswrapper[4948]: E0312 00:08:54.313263 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-f7mp5_openshift-ovn-kubernetes(15117613-9fad-48c7-98c4-a2d84502ded9)\"" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" podUID="15117613-9fad-48c7-98c4-a2d84502ded9" Mar 12 00:08:54 crc kubenswrapper[4948]: I0312 00:08:54.324819 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15117613-9fad-48c7-98c4-a2d84502ded9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4036d7131db00afd8eab8e093626b688297dc776b5af80c3c38b51da7124085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84af4cdd213d90636a889a8a8a3aaca3dee1e2677368efa0ffc768940c2cc994\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e481ace9e84b6aa66cbdda3ee8425c4120a665915d2c70ae4faf60bba6205e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9eadefd0babc119f776d48c9d2c9e4bced58db685af9a8e97f25e6a9a2e3c89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9d5e77753fda70350353daa0967257dd4929dba4440a95495ab4c2a0659aef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cad943edb9a8d445b33a6e70d85b475c1ccba460a6056bccae94fe60fd6b73f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a18330e68c4cbdf0380d4f4845bed125a80893c948f89c674632af8eb1ffb5f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a18330e68c4cbdf0380d4f4845bed125a80893c948f89c674632af8eb1ffb5f4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T00:08:42Z\\\",\\\"message\\\":\\\"lector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0312 00:08:42.392052 7170 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0312 00:08:42.392236 7170 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0312 00:08:42.392701 7170 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0312 00:08:42.392763 7170 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0312 00:08:42.392775 7170 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0312 00:08:42.392819 7170 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0312 00:08:42.392895 7170 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0312 00:08:42.392966 7170 factory.go:656] Stopping watch factory\\\\nI0312 00:08:42.392993 7170 ovnkube.go:599] Stopped ovnkube\\\\nI0312 00:08:42.393027 7170 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0312 00:08:42.393049 7170 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0312 00:08:42.393062 7170 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0312 00:08:42.393074 7170 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0312 00:08:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-f7mp5_openshift-ovn-kubernetes(15117613-9fad-48c7-98c4-a2d84502ded9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46fa6140f4f4bd32e298fa36f6d7af6890029f9bc3080d2b4d94be8307346de8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7mp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:54Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:54 crc kubenswrapper[4948]: I0312 00:08:54.345607 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:08:54Z is after 2025-08-24T17:21:41Z" Mar 12 00:08:56 crc kubenswrapper[4948]: I0312 00:08:56.312480 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:08:56 crc kubenswrapper[4948]: I0312 00:08:56.312544 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:08:56 crc kubenswrapper[4948]: I0312 00:08:56.312553 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:08:56 crc kubenswrapper[4948]: E0312 00:08:56.312768 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 00:08:56 crc kubenswrapper[4948]: I0312 00:08:56.312803 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:08:56 crc kubenswrapper[4948]: E0312 00:08:56.312981 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mqmtd" podUID="be985ff1-de8a-431f-a230-67894ebe52d5" Mar 12 00:08:56 crc kubenswrapper[4948]: E0312 00:08:56.313094 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 00:08:56 crc kubenswrapper[4948]: E0312 00:08:56.313181 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 00:08:56 crc kubenswrapper[4948]: E0312 00:08:56.439223 4948 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 12 00:08:58 crc kubenswrapper[4948]: I0312 00:08:58.312084 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:08:58 crc kubenswrapper[4948]: I0312 00:08:58.312114 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:08:58 crc kubenswrapper[4948]: I0312 00:08:58.312092 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:08:58 crc kubenswrapper[4948]: E0312 00:08:58.312197 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 00:08:58 crc kubenswrapper[4948]: I0312 00:08:58.312492 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:08:58 crc kubenswrapper[4948]: E0312 00:08:58.312489 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 00:08:58 crc kubenswrapper[4948]: E0312 00:08:58.312904 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mqmtd" podUID="be985ff1-de8a-431f-a230-67894ebe52d5" Mar 12 00:08:58 crc kubenswrapper[4948]: E0312 00:08:58.312993 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 00:08:58 crc kubenswrapper[4948]: I0312 00:08:58.326153 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/kube-rbac-proxy-crio-crc"] Mar 12 00:09:00 crc kubenswrapper[4948]: I0312 00:09:00.311532 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:09:00 crc kubenswrapper[4948]: I0312 00:09:00.311593 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:09:00 crc kubenswrapper[4948]: E0312 00:09:00.311712 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 00:09:00 crc kubenswrapper[4948]: I0312 00:09:00.311739 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:09:00 crc kubenswrapper[4948]: I0312 00:09:00.311550 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:09:00 crc kubenswrapper[4948]: E0312 00:09:00.311907 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 00:09:00 crc kubenswrapper[4948]: E0312 00:09:00.312074 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 00:09:00 crc kubenswrapper[4948]: E0312 00:09:00.312210 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mqmtd" podUID="be985ff1-de8a-431f-a230-67894ebe52d5" Mar 12 00:09:01 crc kubenswrapper[4948]: I0312 00:09:01.337194 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfp2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"662dcbe7-7cf8-42d4-9a5e-345c32a49972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff3f3c69dcb28c7b0fd85f9e0c33b3339a1f6159dbf10bcf0efbdce4f7728c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a57d476b5b00110168878bd1c19c32ef122525c0259ac0a70de31c8220996d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4a57d476b5b00110168878bd1c19c32ef122525c0259ac0a70de31c8220996d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bb429b94a1286d8aac77cb77112c7163f7aa50a1c9cab997f8d088c100887c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0bb429b94a1286d8aac77cb77112c7163f7aa50a1c9cab997f8d088c100887c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a77485fd9e246f190724b86663e0dbe81146e1fcc51f468dd05f23c1db878fd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a77485fd9e246f190724b86663e0dbe81146e1fcc51f468dd05f23c1db878fd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adaa20d9a4898057802a8c9759ab3cd485963180bfc4f6b926bf6d2c886bcf27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adaa20d9a4898057802a8c9759ab3cd485963180bfc4f6b926bf6d2c886bcf27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://998fbb934f01c9ecd315f78e776199d4fd538106f8373f09b730d2df23ac9e3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://998fbb934f01c9ecd315f78e776199d4fd538106f8373f09b730d2df23ac9e3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46ebe18804ff07eb0e922a6cc1c39ff93875302d231222642b211e3527fe6373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46ebe18804ff07eb0e922a6cc1c39ff93875302d231222642b211e3527fe6373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfp2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:01Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:01 crc kubenswrapper[4948]: I0312 00:09:01.351985 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e483c2b-08f0-4e92-8e4a-b7281f30af3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d6839c99ab958ad1f40636c70e29bc285292fdbee1e910cdf7f8ab43ca7e8db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd9jv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82de8f2b633f1b6c295149d077bfd2b308f5acaa429acf411ed16f456536d0ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd9jv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4xwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:01Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:01 crc kubenswrapper[4948]: I0312 00:09:01.373557 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c835bb7a-4e06-4047-9fe6-e0c605b36bc6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc9760bd6ed0d092081fdd2ad33dda2d67f1305531640a35b9dce84b5d86a813\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f580535b16c31fb0ebadbdd9b7851ba2dce2f7d52bcad8c9cacc197f5e7dfad4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T00:07:08Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0312 00:06:43.492131 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0312 00:06:43.494442 1 observer_polling.go:159] Starting file observer\\\\nI0312 00:06:43.528056 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0312 00:06:43.536640 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0312 00:07:08.239474 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0312 00:07:08.239540 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb45482a404d6696cfbc3919997e53146b7b9dd1536056a55374574ae01ff193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://663bca87e44109c1d2bb59964e567d8b78fe2e9de39bcec0557cb43e39143add\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bf5b5d40a0cc76725ae2cf453089b31e75c304cf6037ef06b8ca8f7cff35a5a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:06:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:01Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:01 crc kubenswrapper[4948]: I0312 00:09:01.394585 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5139686878eb07d3abced92d0ccfe05be8907580ed75233b72fc2ad3415a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15cbe3fa75b69a7176803295e16a314a8bc73f5797c874124d54e76d47fa1064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:01Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:01 crc kubenswrapper[4948]: I0312 00:09:01.413622 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:01Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:01 crc kubenswrapper[4948]: I0312 00:09:01.428102 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-njc9c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0d6440c-4524-42ea-b18a-0ddc4662f672\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f4d0d9599ad62aaba35ba1414a806f3a163e29a3c34fd893ef5b14eae003591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7wn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-njc9c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:01Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:01 crc kubenswrapper[4948]: E0312 00:09:01.440451 4948 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 12 00:09:01 crc kubenswrapper[4948]: I0312 00:09:01.443835 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6c75836-fbdc-42f6-b581-7d1a827dc688\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f58f45d5eb63c29e243da4ed197ef81a00adb088ef9c8b65323dd3ff49f3c5a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78171d768ed45a0338b479044b9820500c88e17d0d2d9dcbf4b724de06ae3012\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78171d768ed45a0338b479044b9820500c88e17d0d2d9dcbf4b724de06ae3012\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:06:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:01Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:01 crc kubenswrapper[4948]: I0312 00:09:01.467259 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2efff53-8e12-41a1-bcb1-9a1d08c42a4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1cb328c8431b39dac735d49726a2eae40a367849b2c146bb2e49d02958f0e5a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ad13662849131222d5c8ba1041b81fa75bcb1fcbfa2e71c0c751f04b532a5b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8cd4faf47b25395ffd42d4bf5416ea29dcaa3ec380d6d6d6a9f13321d35195a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8028fd7b3e6d26bf30666425c843933b8fb87c3b2a258d510608ec718eb4fe82\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73834c8db23f6536034dce09e38b56665211238280bf9a4e1b13e74d81245a77\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T00:07:50Z\\\",\\\"message\\\":\\\"le observer\\\\nW0312 00:07:49.927227 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0312 00:07:49.927347 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 00:07:49.928000 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2511358960/tls.crt::/tmp/serving-cert-2511358960/tls.key\\\\\\\"\\\\nI0312 00:07:50.321455 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0312 00:07:50.324479 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0312 00:07:50.324511 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0312 00:07:50.324546 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0312 00:07:50.324556 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0312 00:07:50.331591 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0312 00:07:50.331609 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0312 00:07:50.331634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 00:07:50.331642 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 00:07:50.331651 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0312 00:07:50.331658 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0312 00:07:50.331664 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0312 00:07:50.331670 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0312 00:07:50.332917 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:07:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d56e32f9cb96ad103bab80d6a730b70d3387264a25a7a61ef1f0726e7290d6c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e42f2e94463ac8a3d7980a25122062da9ee87fcdc1f10e260ff9974b2e42359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e42f2e94463ac8a3d7980a25122062da9ee87fcdc1f10e260ff9974b2e42359\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:06:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:01Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:01 crc kubenswrapper[4948]: I0312 00:09:01.489370 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5a9ae9441886817563a5fc793bff7c9b8d77d406ceb1d1fb1eebd4307a1c64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:01Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:01 crc kubenswrapper[4948]: I0312 00:09:01.508853 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwpqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b7580c2-5c23-4c67-807a-ea97a3df9398\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d07c603619f858177d4b007e785208853abcfe2dc6c1af435309030fa409010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8lxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwpqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:01Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:01 crc kubenswrapper[4948]: I0312 00:09:01.524746 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a124d35d-f697-49ae-8dac-ee32cd6fc5f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49c25293fe6ddb25b23fc250e3a5d94ce3f2c464f450c0da24dcff2febf15124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a57137a3fb48ab82c4324c6121b7dbcde8dd33e02cd7012ff42248cc5f164bec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccb319a44802931d007a0d9265f98bac6f23efee9a7fb82798718a315464e89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c756f1a5b429d17d95df4797725081682b66e9ef3e53a9deb9518e326e40325b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c756f1a5b429d17d95df4797725081682b66e9ef3e53a9deb9518e326e40325b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:06:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:01Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:01 crc kubenswrapper[4948]: I0312 00:09:01.538075 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9s5fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f9c7f1-6593-4d92-854c-e6c9964a0435\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b8833cbaf3809463e18bfebc9d5bada5fb26823831cc742a3a27177662f60e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xf4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9s5fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:01Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:01 crc kubenswrapper[4948]: I0312 00:09:01.552178 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xdd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b5233d6-d2b8-42dc-b881-18e439ecc442\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0ea55d993c72b6d50a4829ecfff4adae7d0f625a90e40545c9e63201cdda29c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjpwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb23ff4bc484524e306561c0c574e2b2eee39fc80efcd9f44f1924fe42df1a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjpwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4xdd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:01Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:01 crc kubenswrapper[4948]: I0312 00:09:01.580168 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15117613-9fad-48c7-98c4-a2d84502ded9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4036d7131db00afd8eab8e093626b688297dc776b5af80c3c38b51da7124085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84af4cdd213d90636a889a8a8a3aaca3dee1e2677368efa0ffc768940c2cc994\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e481ace9e84b6aa66cbdda3ee8425c4120a665915d2c70ae4faf60bba6205e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9eadefd0babc119f776d48c9d2c9e4bced58db685af9a8e97f25e6a9a2e3c89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9d5e77753fda70350353daa0967257dd4929dba4440a95495ab4c2a0659aef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cad943edb9a8d445b33a6e70d85b475c1ccba460a6056bccae94fe60fd6b73f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a18330e68c4cbdf0380d4f4845bed125a80893c948f89c674632af8eb1ffb5f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a18330e68c4cbdf0380d4f4845bed125a80893c948f89c674632af8eb1ffb5f4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T00:08:42Z\\\",\\\"message\\\":\\\"lector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0312 00:08:42.392052 7170 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0312 00:08:42.392236 7170 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0312 00:08:42.392701 7170 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0312 00:08:42.392763 7170 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0312 00:08:42.392775 7170 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0312 00:08:42.392819 7170 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0312 00:08:42.392895 7170 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0312 00:08:42.392966 7170 factory.go:656] Stopping watch factory\\\\nI0312 00:08:42.392993 7170 ovnkube.go:599] Stopped ovnkube\\\\nI0312 00:08:42.393027 7170 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0312 00:08:42.393049 7170 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0312 00:08:42.393062 7170 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0312 00:08:42.393074 7170 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0312 00:08:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-f7mp5_openshift-ovn-kubernetes(15117613-9fad-48c7-98c4-a2d84502ded9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46fa6140f4f4bd32e298fa36f6d7af6890029f9bc3080d2b4d94be8307346de8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7mp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:01Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:01 crc kubenswrapper[4948]: I0312 00:09:01.597244 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:01Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:01 crc kubenswrapper[4948]: I0312 00:09:01.611005 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:01Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:01 crc kubenswrapper[4948]: I0312 00:09:01.625797 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8d58ee45d75284e030700fa526b2c39197a26f6a391a7417f25a026f356b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:01Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:01 crc kubenswrapper[4948]: I0312 00:09:01.639404 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mqmtd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be985ff1-de8a-431f-a230-67894ebe52d5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmsz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmsz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mqmtd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:01Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:01 crc kubenswrapper[4948]: I0312 00:09:01.840123 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-bwpqp_3b7580c2-5c23-4c67-807a-ea97a3df9398/kube-multus/0.log" Mar 12 00:09:01 crc kubenswrapper[4948]: I0312 00:09:01.840226 4948 generic.go:334] "Generic (PLEG): container finished" podID="3b7580c2-5c23-4c67-807a-ea97a3df9398" containerID="9d07c603619f858177d4b007e785208853abcfe2dc6c1af435309030fa409010" exitCode=1 Mar 12 00:09:01 crc kubenswrapper[4948]: I0312 00:09:01.840286 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-bwpqp" event={"ID":"3b7580c2-5c23-4c67-807a-ea97a3df9398","Type":"ContainerDied","Data":"9d07c603619f858177d4b007e785208853abcfe2dc6c1af435309030fa409010"} Mar 12 00:09:01 crc kubenswrapper[4948]: I0312 00:09:01.841071 4948 scope.go:117] "RemoveContainer" containerID="9d07c603619f858177d4b007e785208853abcfe2dc6c1af435309030fa409010" Mar 12 00:09:01 crc kubenswrapper[4948]: I0312 00:09:01.858823 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwpqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b7580c2-5c23-4c67-807a-ea97a3df9398\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:01Z\\\",\\\"message\\\":\\\"containers with unready status: [kube-multus]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://9d07c603619f858177d4b007e785208853abcfe2dc6c1af435309030fa409010\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d07c603619f858177d4b007e785208853abcfe2dc6c1af435309030fa409010\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T00:09:01Z\\\",\\\"message\\\":\\\"2026-03-12T00:08:16+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_5e485fc1-f819-4bac-9855-3dfbad154bba\\\\n2026-03-12T00:08:16+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_5e485fc1-f819-4bac-9855-3dfbad154bba to /host/opt/cni/bin/\\\\n2026-03-12T00:08:16Z [verbose] multus-daemon started\\\\n2026-03-12T00:08:16Z [verbose] Readiness Indicator file check\\\\n2026-03-12T00:09:01Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8lxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwpqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:01Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:01 crc kubenswrapper[4948]: I0312 00:09:01.882859 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-njc9c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0d6440c-4524-42ea-b18a-0ddc4662f672\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f4d0d9599ad62aaba35ba1414a806f3a163e29a3c34fd893ef5b14eae003591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7wn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-njc9c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:01Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:01 crc kubenswrapper[4948]: I0312 00:09:01.901569 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6c75836-fbdc-42f6-b581-7d1a827dc688\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f58f45d5eb63c29e243da4ed197ef81a00adb088ef9c8b65323dd3ff49f3c5a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78171d768ed45a0338b479044b9820500c88e17d0d2d9dcbf4b724de06ae3012\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78171d768ed45a0338b479044b9820500c88e17d0d2d9dcbf4b724de06ae3012\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:06:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:01Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:01 crc kubenswrapper[4948]: I0312 00:09:01.923244 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2efff53-8e12-41a1-bcb1-9a1d08c42a4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1cb328c8431b39dac735d49726a2eae40a367849b2c146bb2e49d02958f0e5a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ad13662849131222d5c8ba1041b81fa75bcb1fcbfa2e71c0c751f04b532a5b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8cd4faf47b25395ffd42d4bf5416ea29dcaa3ec380d6d6d6a9f13321d35195a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8028fd7b3e6d26bf30666425c843933b8fb87c3b2a258d510608ec718eb4fe82\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73834c8db23f6536034dce09e38b56665211238280bf9a4e1b13e74d81245a77\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T00:07:50Z\\\",\\\"message\\\":\\\"le observer\\\\nW0312 00:07:49.927227 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0312 00:07:49.927347 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 00:07:49.928000 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2511358960/tls.crt::/tmp/serving-cert-2511358960/tls.key\\\\\\\"\\\\nI0312 00:07:50.321455 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0312 00:07:50.324479 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0312 00:07:50.324511 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0312 00:07:50.324546 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0312 00:07:50.324556 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0312 00:07:50.331591 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0312 00:07:50.331609 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0312 00:07:50.331634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 00:07:50.331642 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 00:07:50.331651 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0312 00:07:50.331658 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0312 00:07:50.331664 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0312 00:07:50.331670 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0312 00:07:50.332917 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:07:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d56e32f9cb96ad103bab80d6a730b70d3387264a25a7a61ef1f0726e7290d6c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e42f2e94463ac8a3d7980a25122062da9ee87fcdc1f10e260ff9974b2e42359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e42f2e94463ac8a3d7980a25122062da9ee87fcdc1f10e260ff9974b2e42359\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:06:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:01Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:01 crc kubenswrapper[4948]: I0312 00:09:01.941242 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5a9ae9441886817563a5fc793bff7c9b8d77d406ceb1d1fb1eebd4307a1c64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:01Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:01 crc kubenswrapper[4948]: I0312 00:09:01.957639 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a124d35d-f697-49ae-8dac-ee32cd6fc5f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49c25293fe6ddb25b23fc250e3a5d94ce3f2c464f450c0da24dcff2febf15124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a57137a3fb48ab82c4324c6121b7dbcde8dd33e02cd7012ff42248cc5f164bec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccb319a44802931d007a0d9265f98bac6f23efee9a7fb82798718a315464e89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c756f1a5b429d17d95df4797725081682b66e9ef3e53a9deb9518e326e40325b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c756f1a5b429d17d95df4797725081682b66e9ef3e53a9deb9518e326e40325b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:06:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:01Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:01 crc kubenswrapper[4948]: I0312 00:09:01.971838 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9s5fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f9c7f1-6593-4d92-854c-e6c9964a0435\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b8833cbaf3809463e18bfebc9d5bada5fb26823831cc742a3a27177662f60e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xf4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9s5fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:01Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:01 crc kubenswrapper[4948]: I0312 00:09:01.985774 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xdd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b5233d6-d2b8-42dc-b881-18e439ecc442\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0ea55d993c72b6d50a4829ecfff4adae7d0f625a90e40545c9e63201cdda29c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjpwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb23ff4bc484524e306561c0c574e2b2eee39fc80efcd9f44f1924fe42df1a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjpwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4xdd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:01Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:02 crc kubenswrapper[4948]: I0312 00:09:02.002509 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mqmtd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be985ff1-de8a-431f-a230-67894ebe52d5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmsz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmsz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mqmtd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:02Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:02 crc kubenswrapper[4948]: I0312 00:09:02.031484 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15117613-9fad-48c7-98c4-a2d84502ded9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4036d7131db00afd8eab8e093626b688297dc776b5af80c3c38b51da7124085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84af4cdd213d90636a889a8a8a3aaca3dee1e2677368efa0ffc768940c2cc994\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e481ace9e84b6aa66cbdda3ee8425c4120a665915d2c70ae4faf60bba6205e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9eadefd0babc119f776d48c9d2c9e4bced58db685af9a8e97f25e6a9a2e3c89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9d5e77753fda70350353daa0967257dd4929dba4440a95495ab4c2a0659aef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cad943edb9a8d445b33a6e70d85b475c1ccba460a6056bccae94fe60fd6b73f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a18330e68c4cbdf0380d4f4845bed125a80893c948f89c674632af8eb1ffb5f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a18330e68c4cbdf0380d4f4845bed125a80893c948f89c674632af8eb1ffb5f4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T00:08:42Z\\\",\\\"message\\\":\\\"lector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0312 00:08:42.392052 7170 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0312 00:08:42.392236 7170 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0312 00:08:42.392701 7170 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0312 00:08:42.392763 7170 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0312 00:08:42.392775 7170 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0312 00:08:42.392819 7170 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0312 00:08:42.392895 7170 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0312 00:08:42.392966 7170 factory.go:656] Stopping watch factory\\\\nI0312 00:08:42.392993 7170 ovnkube.go:599] Stopped ovnkube\\\\nI0312 00:08:42.393027 7170 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0312 00:08:42.393049 7170 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0312 00:08:42.393062 7170 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0312 00:08:42.393074 7170 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0312 00:08:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-f7mp5_openshift-ovn-kubernetes(15117613-9fad-48c7-98c4-a2d84502ded9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46fa6140f4f4bd32e298fa36f6d7af6890029f9bc3080d2b4d94be8307346de8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7mp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:02Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:02 crc kubenswrapper[4948]: I0312 00:09:02.051349 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:02Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:02 crc kubenswrapper[4948]: I0312 00:09:02.073438 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:02Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:02 crc kubenswrapper[4948]: I0312 00:09:02.093936 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8d58ee45d75284e030700fa526b2c39197a26f6a391a7417f25a026f356b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:02Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:02 crc kubenswrapper[4948]: I0312 00:09:02.142444 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:02Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:02 crc kubenswrapper[4948]: I0312 00:09:02.166677 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfp2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"662dcbe7-7cf8-42d4-9a5e-345c32a49972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff3f3c69dcb28c7b0fd85f9e0c33b3339a1f6159dbf10bcf0efbdce4f7728c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a57d476b5b00110168878bd1c19c32ef122525c0259ac0a70de31c8220996d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4a57d476b5b00110168878bd1c19c32ef122525c0259ac0a70de31c8220996d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bb429b94a1286d8aac77cb77112c7163f7aa50a1c9cab997f8d088c100887c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0bb429b94a1286d8aac77cb77112c7163f7aa50a1c9cab997f8d088c100887c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a77485fd9e246f190724b86663e0dbe81146e1fcc51f468dd05f23c1db878fd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a77485fd9e246f190724b86663e0dbe81146e1fcc51f468dd05f23c1db878fd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adaa20d9a4898057802a8c9759ab3cd485963180bfc4f6b926bf6d2c886bcf27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adaa20d9a4898057802a8c9759ab3cd485963180bfc4f6b926bf6d2c886bcf27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://998fbb934f01c9ecd315f78e776199d4fd538106f8373f09b730d2df23ac9e3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://998fbb934f01c9ecd315f78e776199d4fd538106f8373f09b730d2df23ac9e3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46ebe18804ff07eb0e922a6cc1c39ff93875302d231222642b211e3527fe6373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46ebe18804ff07eb0e922a6cc1c39ff93875302d231222642b211e3527fe6373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfp2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:02Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:02 crc kubenswrapper[4948]: I0312 00:09:02.216540 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e483c2b-08f0-4e92-8e4a-b7281f30af3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d6839c99ab958ad1f40636c70e29bc285292fdbee1e910cdf7f8ab43ca7e8db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd9jv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82de8f2b633f1b6c295149d077bfd2b308f5acaa429acf411ed16f456536d0ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd9jv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4xwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:02Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:02 crc kubenswrapper[4948]: I0312 00:09:02.241699 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c835bb7a-4e06-4047-9fe6-e0c605b36bc6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc9760bd6ed0d092081fdd2ad33dda2d67f1305531640a35b9dce84b5d86a813\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f580535b16c31fb0ebadbdd9b7851ba2dce2f7d52bcad8c9cacc197f5e7dfad4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T00:07:08Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0312 00:06:43.492131 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0312 00:06:43.494442 1 observer_polling.go:159] Starting file observer\\\\nI0312 00:06:43.528056 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0312 00:06:43.536640 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0312 00:07:08.239474 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0312 00:07:08.239540 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb45482a404d6696cfbc3919997e53146b7b9dd1536056a55374574ae01ff193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://663bca87e44109c1d2bb59964e567d8b78fe2e9de39bcec0557cb43e39143add\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bf5b5d40a0cc76725ae2cf453089b31e75c304cf6037ef06b8ca8f7cff35a5a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:06:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:02Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:02 crc kubenswrapper[4948]: I0312 00:09:02.260924 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5139686878eb07d3abced92d0ccfe05be8907580ed75233b72fc2ad3415a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15cbe3fa75b69a7176803295e16a314a8bc73f5797c874124d54e76d47fa1064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:02Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:02 crc kubenswrapper[4948]: I0312 00:09:02.312465 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:09:02 crc kubenswrapper[4948]: I0312 00:09:02.312541 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:09:02 crc kubenswrapper[4948]: I0312 00:09:02.312542 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:09:02 crc kubenswrapper[4948]: I0312 00:09:02.312474 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:09:02 crc kubenswrapper[4948]: E0312 00:09:02.312628 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 00:09:02 crc kubenswrapper[4948]: E0312 00:09:02.312732 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mqmtd" podUID="be985ff1-de8a-431f-a230-67894ebe52d5" Mar 12 00:09:02 crc kubenswrapper[4948]: E0312 00:09:02.312831 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 00:09:02 crc kubenswrapper[4948]: E0312 00:09:02.312942 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 00:09:02 crc kubenswrapper[4948]: I0312 00:09:02.847354 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-bwpqp_3b7580c2-5c23-4c67-807a-ea97a3df9398/kube-multus/0.log" Mar 12 00:09:02 crc kubenswrapper[4948]: I0312 00:09:02.848469 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-bwpqp" event={"ID":"3b7580c2-5c23-4c67-807a-ea97a3df9398","Type":"ContainerStarted","Data":"b28502ee0add02df45488b29f20d8be4102b0f3faaf8b5e26b5943a410163bf7"} Mar 12 00:09:02 crc kubenswrapper[4948]: I0312 00:09:02.855170 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:09:02 crc kubenswrapper[4948]: I0312 00:09:02.855233 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:09:02 crc kubenswrapper[4948]: I0312 00:09:02.855251 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:09:02 crc kubenswrapper[4948]: I0312 00:09:02.855278 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:09:02 crc kubenswrapper[4948]: I0312 00:09:02.855294 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:09:02Z","lastTransitionTime":"2026-03-12T00:09:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:09:02 crc kubenswrapper[4948]: I0312 00:09:02.864462 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9s5fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f9c7f1-6593-4d92-854c-e6c9964a0435\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b8833cbaf3809463e18bfebc9d5bada5fb26823831cc742a3a27177662f60e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xf4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9s5fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:02Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:02 crc kubenswrapper[4948]: E0312 00:09:02.883054 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:09:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:09:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:09:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:09:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17825588-29cf-4059-bab7-1b042b9bf1f3\\\",\\\"systemUUID\\\":\\\"b18e4bf2-d37e-4b80-8da3-cd3803754ba2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:02Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:02 crc kubenswrapper[4948]: I0312 00:09:02.887860 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xdd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b5233d6-d2b8-42dc-b881-18e439ecc442\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0ea55d993c72b6d50a4829ecfff4adae7d0f625a90e40545c9e63201cdda29c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjpwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb23ff4bc484524e306561c0c574e2b2eee39fc80efcd9f44f1924fe42df1a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjpwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4xdd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:02Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:02 crc kubenswrapper[4948]: I0312 00:09:02.889660 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:09:02 crc kubenswrapper[4948]: I0312 00:09:02.889731 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:09:02 crc kubenswrapper[4948]: I0312 00:09:02.889750 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:09:02 crc kubenswrapper[4948]: I0312 00:09:02.889774 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:09:02 crc kubenswrapper[4948]: I0312 00:09:02.889791 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:09:02Z","lastTransitionTime":"2026-03-12T00:09:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:09:02 crc kubenswrapper[4948]: I0312 00:09:02.910757 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a124d35d-f697-49ae-8dac-ee32cd6fc5f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49c25293fe6ddb25b23fc250e3a5d94ce3f2c464f450c0da24dcff2febf15124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a57137a3fb48ab82c4324c6121b7dbcde8dd33e02cd7012ff42248cc5f164bec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccb319a44802931d007a0d9265f98bac6f23efee9a7fb82798718a315464e89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c756f1a5b429d17d95df4797725081682b66e9ef3e53a9deb9518e326e40325b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c756f1a5b429d17d95df4797725081682b66e9ef3e53a9deb9518e326e40325b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:06:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:02Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:02 crc kubenswrapper[4948]: E0312 00:09:02.914668 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:09:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:09:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:09:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:09:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17825588-29cf-4059-bab7-1b042b9bf1f3\\\",\\\"systemUUID\\\":\\\"b18e4bf2-d37e-4b80-8da3-cd3803754ba2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:02Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:02 crc kubenswrapper[4948]: I0312 00:09:02.921468 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:09:02 crc kubenswrapper[4948]: I0312 00:09:02.921531 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:09:02 crc kubenswrapper[4948]: I0312 00:09:02.921548 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:09:02 crc kubenswrapper[4948]: I0312 00:09:02.921576 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:09:02 crc kubenswrapper[4948]: I0312 00:09:02.921594 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:09:02Z","lastTransitionTime":"2026-03-12T00:09:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:09:02 crc kubenswrapper[4948]: I0312 00:09:02.928615 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:02Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:02 crc kubenswrapper[4948]: E0312 00:09:02.941819 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:09:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:09:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:09:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:09:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17825588-29cf-4059-bab7-1b042b9bf1f3\\\",\\\"systemUUID\\\":\\\"b18e4bf2-d37e-4b80-8da3-cd3803754ba2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:02Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:02 crc kubenswrapper[4948]: I0312 00:09:02.946151 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8d58ee45d75284e030700fa526b2c39197a26f6a391a7417f25a026f356b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:02Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:02 crc kubenswrapper[4948]: I0312 00:09:02.946764 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:09:02 crc kubenswrapper[4948]: I0312 00:09:02.946813 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:09:02 crc kubenswrapper[4948]: I0312 00:09:02.946830 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:09:02 crc kubenswrapper[4948]: I0312 00:09:02.946854 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:09:02 crc kubenswrapper[4948]: I0312 00:09:02.946871 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:09:02Z","lastTransitionTime":"2026-03-12T00:09:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:09:02 crc kubenswrapper[4948]: I0312 00:09:02.961625 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mqmtd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be985ff1-de8a-431f-a230-67894ebe52d5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmsz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmsz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mqmtd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:02Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:02 crc kubenswrapper[4948]: E0312 00:09:02.965878 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:09:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:09:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:09:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:09:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17825588-29cf-4059-bab7-1b042b9bf1f3\\\",\\\"systemUUID\\\":\\\"b18e4bf2-d37e-4b80-8da3-cd3803754ba2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:02Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:02 crc kubenswrapper[4948]: I0312 00:09:02.970868 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:09:02 crc kubenswrapper[4948]: I0312 00:09:02.970918 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:09:02 crc kubenswrapper[4948]: I0312 00:09:02.970939 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:09:02 crc kubenswrapper[4948]: I0312 00:09:02.970961 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:09:02 crc kubenswrapper[4948]: I0312 00:09:02.970979 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:09:02Z","lastTransitionTime":"2026-03-12T00:09:02Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:09:02 crc kubenswrapper[4948]: E0312 00:09:02.989640 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:09:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:09:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:02Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:09:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:02Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:09:02Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:02Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17825588-29cf-4059-bab7-1b042b9bf1f3\\\",\\\"systemUUID\\\":\\\"b18e4bf2-d37e-4b80-8da3-cd3803754ba2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:02Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:02 crc kubenswrapper[4948]: E0312 00:09:02.989918 4948 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 12 00:09:02 crc kubenswrapper[4948]: I0312 00:09:02.991255 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15117613-9fad-48c7-98c4-a2d84502ded9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4036d7131db00afd8eab8e093626b688297dc776b5af80c3c38b51da7124085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84af4cdd213d90636a889a8a8a3aaca3dee1e2677368efa0ffc768940c2cc994\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e481ace9e84b6aa66cbdda3ee8425c4120a665915d2c70ae4faf60bba6205e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9eadefd0babc119f776d48c9d2c9e4bced58db685af9a8e97f25e6a9a2e3c89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9d5e77753fda70350353daa0967257dd4929dba4440a95495ab4c2a0659aef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cad943edb9a8d445b33a6e70d85b475c1ccba460a6056bccae94fe60fd6b73f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a18330e68c4cbdf0380d4f4845bed125a80893c948f89c674632af8eb1ffb5f4\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a18330e68c4cbdf0380d4f4845bed125a80893c948f89c674632af8eb1ffb5f4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T00:08:42Z\\\",\\\"message\\\":\\\"lector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0312 00:08:42.392052 7170 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0312 00:08:42.392236 7170 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0312 00:08:42.392701 7170 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0312 00:08:42.392763 7170 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0312 00:08:42.392775 7170 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0312 00:08:42.392819 7170 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0312 00:08:42.392895 7170 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0312 00:08:42.392966 7170 factory.go:656] Stopping watch factory\\\\nI0312 00:08:42.392993 7170 ovnkube.go:599] Stopped ovnkube\\\\nI0312 00:08:42.393027 7170 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0312 00:08:42.393049 7170 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0312 00:08:42.393062 7170 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0312 00:08:42.393074 7170 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0312 00:08:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":2,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 20s restarting failed container=ovnkube-controller pod=ovnkube-node-f7mp5_openshift-ovn-kubernetes(15117613-9fad-48c7-98c4-a2d84502ded9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46fa6140f4f4bd32e298fa36f6d7af6890029f9bc3080d2b4d94be8307346de8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7mp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:02Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:03 crc kubenswrapper[4948]: I0312 00:09:03.009784 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:03Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:03 crc kubenswrapper[4948]: I0312 00:09:03.028865 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5139686878eb07d3abced92d0ccfe05be8907580ed75233b72fc2ad3415a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15cbe3fa75b69a7176803295e16a314a8bc73f5797c874124d54e76d47fa1064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:03Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:03 crc kubenswrapper[4948]: I0312 00:09:03.049049 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:03Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:03 crc kubenswrapper[4948]: I0312 00:09:03.070125 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfp2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"662dcbe7-7cf8-42d4-9a5e-345c32a49972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff3f3c69dcb28c7b0fd85f9e0c33b3339a1f6159dbf10bcf0efbdce4f7728c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a57d476b5b00110168878bd1c19c32ef122525c0259ac0a70de31c8220996d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4a57d476b5b00110168878bd1c19c32ef122525c0259ac0a70de31c8220996d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bb429b94a1286d8aac77cb77112c7163f7aa50a1c9cab997f8d088c100887c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0bb429b94a1286d8aac77cb77112c7163f7aa50a1c9cab997f8d088c100887c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a77485fd9e246f190724b86663e0dbe81146e1fcc51f468dd05f23c1db878fd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a77485fd9e246f190724b86663e0dbe81146e1fcc51f468dd05f23c1db878fd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adaa20d9a4898057802a8c9759ab3cd485963180bfc4f6b926bf6d2c886bcf27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adaa20d9a4898057802a8c9759ab3cd485963180bfc4f6b926bf6d2c886bcf27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://998fbb934f01c9ecd315f78e776199d4fd538106f8373f09b730d2df23ac9e3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://998fbb934f01c9ecd315f78e776199d4fd538106f8373f09b730d2df23ac9e3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46ebe18804ff07eb0e922a6cc1c39ff93875302d231222642b211e3527fe6373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46ebe18804ff07eb0e922a6cc1c39ff93875302d231222642b211e3527fe6373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfp2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:03Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:03 crc kubenswrapper[4948]: I0312 00:09:03.087814 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e483c2b-08f0-4e92-8e4a-b7281f30af3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d6839c99ab958ad1f40636c70e29bc285292fdbee1e910cdf7f8ab43ca7e8db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd9jv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82de8f2b633f1b6c295149d077bfd2b308f5acaa429acf411ed16f456536d0ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd9jv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4xwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:03Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:03 crc kubenswrapper[4948]: I0312 00:09:03.106293 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c835bb7a-4e06-4047-9fe6-e0c605b36bc6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc9760bd6ed0d092081fdd2ad33dda2d67f1305531640a35b9dce84b5d86a813\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f580535b16c31fb0ebadbdd9b7851ba2dce2f7d52bcad8c9cacc197f5e7dfad4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T00:07:08Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0312 00:06:43.492131 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0312 00:06:43.494442 1 observer_polling.go:159] Starting file observer\\\\nI0312 00:06:43.528056 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0312 00:06:43.536640 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0312 00:07:08.239474 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0312 00:07:08.239540 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb45482a404d6696cfbc3919997e53146b7b9dd1536056a55374574ae01ff193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://663bca87e44109c1d2bb59964e567d8b78fe2e9de39bcec0557cb43e39143add\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bf5b5d40a0cc76725ae2cf453089b31e75c304cf6037ef06b8ca8f7cff35a5a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:06:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:03Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:03 crc kubenswrapper[4948]: I0312 00:09:03.125695 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2efff53-8e12-41a1-bcb1-9a1d08c42a4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1cb328c8431b39dac735d49726a2eae40a367849b2c146bb2e49d02958f0e5a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ad13662849131222d5c8ba1041b81fa75bcb1fcbfa2e71c0c751f04b532a5b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8cd4faf47b25395ffd42d4bf5416ea29dcaa3ec380d6d6d6a9f13321d35195a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8028fd7b3e6d26bf30666425c843933b8fb87c3b2a258d510608ec718eb4fe82\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73834c8db23f6536034dce09e38b56665211238280bf9a4e1b13e74d81245a77\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T00:07:50Z\\\",\\\"message\\\":\\\"le observer\\\\nW0312 00:07:49.927227 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0312 00:07:49.927347 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 00:07:49.928000 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2511358960/tls.crt::/tmp/serving-cert-2511358960/tls.key\\\\\\\"\\\\nI0312 00:07:50.321455 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0312 00:07:50.324479 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0312 00:07:50.324511 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0312 00:07:50.324546 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0312 00:07:50.324556 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0312 00:07:50.331591 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0312 00:07:50.331609 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0312 00:07:50.331634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 00:07:50.331642 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 00:07:50.331651 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0312 00:07:50.331658 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0312 00:07:50.331664 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0312 00:07:50.331670 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0312 00:07:50.332917 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:07:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d56e32f9cb96ad103bab80d6a730b70d3387264a25a7a61ef1f0726e7290d6c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e42f2e94463ac8a3d7980a25122062da9ee87fcdc1f10e260ff9974b2e42359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e42f2e94463ac8a3d7980a25122062da9ee87fcdc1f10e260ff9974b2e42359\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:06:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:03Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:03 crc kubenswrapper[4948]: I0312 00:09:03.144110 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5a9ae9441886817563a5fc793bff7c9b8d77d406ceb1d1fb1eebd4307a1c64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:03Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:03 crc kubenswrapper[4948]: I0312 00:09:03.164225 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwpqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b7580c2-5c23-4c67-807a-ea97a3df9398\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b28502ee0add02df45488b29f20d8be4102b0f3faaf8b5e26b5943a410163bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d07c603619f858177d4b007e785208853abcfe2dc6c1af435309030fa409010\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T00:09:01Z\\\",\\\"message\\\":\\\"2026-03-12T00:08:16+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_5e485fc1-f819-4bac-9855-3dfbad154bba\\\\n2026-03-12T00:08:16+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_5e485fc1-f819-4bac-9855-3dfbad154bba to /host/opt/cni/bin/\\\\n2026-03-12T00:08:16Z [verbose] multus-daemon started\\\\n2026-03-12T00:08:16Z [verbose] Readiness Indicator file check\\\\n2026-03-12T00:09:01Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8lxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwpqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:03Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:03 crc kubenswrapper[4948]: I0312 00:09:03.178593 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-njc9c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0d6440c-4524-42ea-b18a-0ddc4662f672\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f4d0d9599ad62aaba35ba1414a806f3a163e29a3c34fd893ef5b14eae003591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7wn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-njc9c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:03Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:03 crc kubenswrapper[4948]: I0312 00:09:03.191861 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6c75836-fbdc-42f6-b581-7d1a827dc688\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f58f45d5eb63c29e243da4ed197ef81a00adb088ef9c8b65323dd3ff49f3c5a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78171d768ed45a0338b479044b9820500c88e17d0d2d9dcbf4b724de06ae3012\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78171d768ed45a0338b479044b9820500c88e17d0d2d9dcbf4b724de06ae3012\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:06:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:03Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:04 crc kubenswrapper[4948]: I0312 00:09:04.311636 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:09:04 crc kubenswrapper[4948]: I0312 00:09:04.311710 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:09:04 crc kubenswrapper[4948]: I0312 00:09:04.311756 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:09:04 crc kubenswrapper[4948]: I0312 00:09:04.311828 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:09:04 crc kubenswrapper[4948]: E0312 00:09:04.311826 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mqmtd" podUID="be985ff1-de8a-431f-a230-67894ebe52d5" Mar 12 00:09:04 crc kubenswrapper[4948]: E0312 00:09:04.312062 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 00:09:04 crc kubenswrapper[4948]: E0312 00:09:04.312175 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 00:09:04 crc kubenswrapper[4948]: E0312 00:09:04.312256 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 00:09:06 crc kubenswrapper[4948]: I0312 00:09:06.311804 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:09:06 crc kubenswrapper[4948]: I0312 00:09:06.311836 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:09:06 crc kubenswrapper[4948]: I0312 00:09:06.312272 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:09:06 crc kubenswrapper[4948]: I0312 00:09:06.312288 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:09:06 crc kubenswrapper[4948]: E0312 00:09:06.312450 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 00:09:06 crc kubenswrapper[4948]: E0312 00:09:06.312540 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mqmtd" podUID="be985ff1-de8a-431f-a230-67894ebe52d5" Mar 12 00:09:06 crc kubenswrapper[4948]: E0312 00:09:06.312786 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 00:09:06 crc kubenswrapper[4948]: E0312 00:09:06.312851 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 00:09:06 crc kubenswrapper[4948]: I0312 00:09:06.312928 4948 scope.go:117] "RemoveContainer" containerID="a18330e68c4cbdf0380d4f4845bed125a80893c948f89c674632af8eb1ffb5f4" Mar 12 00:09:06 crc kubenswrapper[4948]: E0312 00:09:06.442241 4948 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 12 00:09:06 crc kubenswrapper[4948]: I0312 00:09:06.864568 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f7mp5_15117613-9fad-48c7-98c4-a2d84502ded9/ovnkube-controller/2.log" Mar 12 00:09:06 crc kubenswrapper[4948]: I0312 00:09:06.867733 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" event={"ID":"15117613-9fad-48c7-98c4-a2d84502ded9","Type":"ContainerStarted","Data":"1d6a0a3fceb31bd31a6c978f0b3ac7ff768469ab3f311918e9dc72b424b00c27"} Mar 12 00:09:06 crc kubenswrapper[4948]: I0312 00:09:06.868422 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:09:06 crc kubenswrapper[4948]: I0312 00:09:06.883293 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6c75836-fbdc-42f6-b581-7d1a827dc688\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f58f45d5eb63c29e243da4ed197ef81a00adb088ef9c8b65323dd3ff49f3c5a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78171d768ed45a0338b479044b9820500c88e17d0d2d9dcbf4b724de06ae3012\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78171d768ed45a0338b479044b9820500c88e17d0d2d9dcbf4b724de06ae3012\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:06:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:06Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:06 crc kubenswrapper[4948]: I0312 00:09:06.900092 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2efff53-8e12-41a1-bcb1-9a1d08c42a4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1cb328c8431b39dac735d49726a2eae40a367849b2c146bb2e49d02958f0e5a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ad13662849131222d5c8ba1041b81fa75bcb1fcbfa2e71c0c751f04b532a5b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8cd4faf47b25395ffd42d4bf5416ea29dcaa3ec380d6d6d6a9f13321d35195a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8028fd7b3e6d26bf30666425c843933b8fb87c3b2a258d510608ec718eb4fe82\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73834c8db23f6536034dce09e38b56665211238280bf9a4e1b13e74d81245a77\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T00:07:50Z\\\",\\\"message\\\":\\\"le observer\\\\nW0312 00:07:49.927227 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0312 00:07:49.927347 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 00:07:49.928000 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2511358960/tls.crt::/tmp/serving-cert-2511358960/tls.key\\\\\\\"\\\\nI0312 00:07:50.321455 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0312 00:07:50.324479 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0312 00:07:50.324511 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0312 00:07:50.324546 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0312 00:07:50.324556 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0312 00:07:50.331591 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0312 00:07:50.331609 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0312 00:07:50.331634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 00:07:50.331642 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 00:07:50.331651 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0312 00:07:50.331658 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0312 00:07:50.331664 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0312 00:07:50.331670 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0312 00:07:50.332917 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:07:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d56e32f9cb96ad103bab80d6a730b70d3387264a25a7a61ef1f0726e7290d6c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e42f2e94463ac8a3d7980a25122062da9ee87fcdc1f10e260ff9974b2e42359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e42f2e94463ac8a3d7980a25122062da9ee87fcdc1f10e260ff9974b2e42359\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:06:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:06Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:06 crc kubenswrapper[4948]: I0312 00:09:06.915900 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5a9ae9441886817563a5fc793bff7c9b8d77d406ceb1d1fb1eebd4307a1c64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:06Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:06 crc kubenswrapper[4948]: I0312 00:09:06.933386 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwpqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b7580c2-5c23-4c67-807a-ea97a3df9398\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b28502ee0add02df45488b29f20d8be4102b0f3faaf8b5e26b5943a410163bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d07c603619f858177d4b007e785208853abcfe2dc6c1af435309030fa409010\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T00:09:01Z\\\",\\\"message\\\":\\\"2026-03-12T00:08:16+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_5e485fc1-f819-4bac-9855-3dfbad154bba\\\\n2026-03-12T00:08:16+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_5e485fc1-f819-4bac-9855-3dfbad154bba to /host/opt/cni/bin/\\\\n2026-03-12T00:08:16Z [verbose] multus-daemon started\\\\n2026-03-12T00:08:16Z [verbose] Readiness Indicator file check\\\\n2026-03-12T00:09:01Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8lxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwpqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:06Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:06 crc kubenswrapper[4948]: I0312 00:09:06.946234 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-njc9c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0d6440c-4524-42ea-b18a-0ddc4662f672\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f4d0d9599ad62aaba35ba1414a806f3a163e29a3c34fd893ef5b14eae003591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7wn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-njc9c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:06Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:06 crc kubenswrapper[4948]: I0312 00:09:06.959118 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a124d35d-f697-49ae-8dac-ee32cd6fc5f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49c25293fe6ddb25b23fc250e3a5d94ce3f2c464f450c0da24dcff2febf15124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a57137a3fb48ab82c4324c6121b7dbcde8dd33e02cd7012ff42248cc5f164bec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccb319a44802931d007a0d9265f98bac6f23efee9a7fb82798718a315464e89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c756f1a5b429d17d95df4797725081682b66e9ef3e53a9deb9518e326e40325b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c756f1a5b429d17d95df4797725081682b66e9ef3e53a9deb9518e326e40325b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:06:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:06Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:06 crc kubenswrapper[4948]: I0312 00:09:06.972629 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9s5fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f9c7f1-6593-4d92-854c-e6c9964a0435\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b8833cbaf3809463e18bfebc9d5bada5fb26823831cc742a3a27177662f60e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xf4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9s5fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:06Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:06 crc kubenswrapper[4948]: I0312 00:09:06.989083 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xdd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b5233d6-d2b8-42dc-b881-18e439ecc442\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0ea55d993c72b6d50a4829ecfff4adae7d0f625a90e40545c9e63201cdda29c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjpwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb23ff4bc484524e306561c0c574e2b2eee39fc80efcd9f44f1924fe42df1a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjpwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4xdd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:06Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:07 crc kubenswrapper[4948]: I0312 00:09:07.009483 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:07Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:07 crc kubenswrapper[4948]: I0312 00:09:07.024252 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:07Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:07 crc kubenswrapper[4948]: I0312 00:09:07.032870 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8d58ee45d75284e030700fa526b2c39197a26f6a391a7417f25a026f356b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:07Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:07 crc kubenswrapper[4948]: I0312 00:09:07.040595 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mqmtd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be985ff1-de8a-431f-a230-67894ebe52d5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmsz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmsz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mqmtd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:07Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:07 crc kubenswrapper[4948]: I0312 00:09:07.059536 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15117613-9fad-48c7-98c4-a2d84502ded9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4036d7131db00afd8eab8e093626b688297dc776b5af80c3c38b51da7124085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84af4cdd213d90636a889a8a8a3aaca3dee1e2677368efa0ffc768940c2cc994\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e481ace9e84b6aa66cbdda3ee8425c4120a665915d2c70ae4faf60bba6205e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9eadefd0babc119f776d48c9d2c9e4bced58db685af9a8e97f25e6a9a2e3c89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9d5e77753fda70350353daa0967257dd4929dba4440a95495ab4c2a0659aef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cad943edb9a8d445b33a6e70d85b475c1ccba460a6056bccae94fe60fd6b73f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d6a0a3fceb31bd31a6c978f0b3ac7ff768469ab3f311918e9dc72b424b00c27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a18330e68c4cbdf0380d4f4845bed125a80893c948f89c674632af8eb1ffb5f4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T00:08:42Z\\\",\\\"message\\\":\\\"lector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0312 00:08:42.392052 7170 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0312 00:08:42.392236 7170 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0312 00:08:42.392701 7170 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0312 00:08:42.392763 7170 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0312 00:08:42.392775 7170 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0312 00:08:42.392819 7170 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0312 00:08:42.392895 7170 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0312 00:08:42.392966 7170 factory.go:656] Stopping watch factory\\\\nI0312 00:08:42.392993 7170 ovnkube.go:599] Stopped ovnkube\\\\nI0312 00:08:42.393027 7170 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0312 00:08:42.393049 7170 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0312 00:08:42.393062 7170 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0312 00:08:42.393074 7170 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0312 00:08:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46fa6140f4f4bd32e298fa36f6d7af6890029f9bc3080d2b4d94be8307346de8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7mp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:07Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:07 crc kubenswrapper[4948]: I0312 00:09:07.069367 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c835bb7a-4e06-4047-9fe6-e0c605b36bc6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc9760bd6ed0d092081fdd2ad33dda2d67f1305531640a35b9dce84b5d86a813\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f580535b16c31fb0ebadbdd9b7851ba2dce2f7d52bcad8c9cacc197f5e7dfad4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T00:07:08Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0312 00:06:43.492131 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0312 00:06:43.494442 1 observer_polling.go:159] Starting file observer\\\\nI0312 00:06:43.528056 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0312 00:06:43.536640 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0312 00:07:08.239474 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0312 00:07:08.239540 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb45482a404d6696cfbc3919997e53146b7b9dd1536056a55374574ae01ff193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://663bca87e44109c1d2bb59964e567d8b78fe2e9de39bcec0557cb43e39143add\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bf5b5d40a0cc76725ae2cf453089b31e75c304cf6037ef06b8ca8f7cff35a5a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:06:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:07Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:07 crc kubenswrapper[4948]: I0312 00:09:07.081158 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5139686878eb07d3abced92d0ccfe05be8907580ed75233b72fc2ad3415a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15cbe3fa75b69a7176803295e16a314a8bc73f5797c874124d54e76d47fa1064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:07Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:07 crc kubenswrapper[4948]: I0312 00:09:07.091399 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:07Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:07 crc kubenswrapper[4948]: I0312 00:09:07.104559 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfp2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"662dcbe7-7cf8-42d4-9a5e-345c32a49972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff3f3c69dcb28c7b0fd85f9e0c33b3339a1f6159dbf10bcf0efbdce4f7728c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a57d476b5b00110168878bd1c19c32ef122525c0259ac0a70de31c8220996d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4a57d476b5b00110168878bd1c19c32ef122525c0259ac0a70de31c8220996d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bb429b94a1286d8aac77cb77112c7163f7aa50a1c9cab997f8d088c100887c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0bb429b94a1286d8aac77cb77112c7163f7aa50a1c9cab997f8d088c100887c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a77485fd9e246f190724b86663e0dbe81146e1fcc51f468dd05f23c1db878fd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a77485fd9e246f190724b86663e0dbe81146e1fcc51f468dd05f23c1db878fd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adaa20d9a4898057802a8c9759ab3cd485963180bfc4f6b926bf6d2c886bcf27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adaa20d9a4898057802a8c9759ab3cd485963180bfc4f6b926bf6d2c886bcf27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://998fbb934f01c9ecd315f78e776199d4fd538106f8373f09b730d2df23ac9e3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://998fbb934f01c9ecd315f78e776199d4fd538106f8373f09b730d2df23ac9e3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46ebe18804ff07eb0e922a6cc1c39ff93875302d231222642b211e3527fe6373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46ebe18804ff07eb0e922a6cc1c39ff93875302d231222642b211e3527fe6373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfp2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:07Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:07 crc kubenswrapper[4948]: I0312 00:09:07.116234 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e483c2b-08f0-4e92-8e4a-b7281f30af3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d6839c99ab958ad1f40636c70e29bc285292fdbee1e910cdf7f8ab43ca7e8db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd9jv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82de8f2b633f1b6c295149d077bfd2b308f5acaa429acf411ed16f456536d0ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd9jv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4xwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:07Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:07 crc kubenswrapper[4948]: I0312 00:09:07.874900 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f7mp5_15117613-9fad-48c7-98c4-a2d84502ded9/ovnkube-controller/3.log" Mar 12 00:09:07 crc kubenswrapper[4948]: I0312 00:09:07.875934 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f7mp5_15117613-9fad-48c7-98c4-a2d84502ded9/ovnkube-controller/2.log" Mar 12 00:09:07 crc kubenswrapper[4948]: I0312 00:09:07.879510 4948 generic.go:334] "Generic (PLEG): container finished" podID="15117613-9fad-48c7-98c4-a2d84502ded9" containerID="1d6a0a3fceb31bd31a6c978f0b3ac7ff768469ab3f311918e9dc72b424b00c27" exitCode=1 Mar 12 00:09:07 crc kubenswrapper[4948]: I0312 00:09:07.879563 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" event={"ID":"15117613-9fad-48c7-98c4-a2d84502ded9","Type":"ContainerDied","Data":"1d6a0a3fceb31bd31a6c978f0b3ac7ff768469ab3f311918e9dc72b424b00c27"} Mar 12 00:09:07 crc kubenswrapper[4948]: I0312 00:09:07.879604 4948 scope.go:117] "RemoveContainer" containerID="a18330e68c4cbdf0380d4f4845bed125a80893c948f89c674632af8eb1ffb5f4" Mar 12 00:09:07 crc kubenswrapper[4948]: I0312 00:09:07.881390 4948 scope.go:117] "RemoveContainer" containerID="1d6a0a3fceb31bd31a6c978f0b3ac7ff768469ab3f311918e9dc72b424b00c27" Mar 12 00:09:07 crc kubenswrapper[4948]: E0312 00:09:07.881705 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-f7mp5_openshift-ovn-kubernetes(15117613-9fad-48c7-98c4-a2d84502ded9)\"" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" podUID="15117613-9fad-48c7-98c4-a2d84502ded9" Mar 12 00:09:07 crc kubenswrapper[4948]: I0312 00:09:07.897979 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5139686878eb07d3abced92d0ccfe05be8907580ed75233b72fc2ad3415a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15cbe3fa75b69a7176803295e16a314a8bc73f5797c874124d54e76d47fa1064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:07Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:07 crc kubenswrapper[4948]: I0312 00:09:07.917368 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:07Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:07 crc kubenswrapper[4948]: I0312 00:09:07.940258 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfp2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"662dcbe7-7cf8-42d4-9a5e-345c32a49972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff3f3c69dcb28c7b0fd85f9e0c33b3339a1f6159dbf10bcf0efbdce4f7728c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a57d476b5b00110168878bd1c19c32ef122525c0259ac0a70de31c8220996d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4a57d476b5b00110168878bd1c19c32ef122525c0259ac0a70de31c8220996d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bb429b94a1286d8aac77cb77112c7163f7aa50a1c9cab997f8d088c100887c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0bb429b94a1286d8aac77cb77112c7163f7aa50a1c9cab997f8d088c100887c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a77485fd9e246f190724b86663e0dbe81146e1fcc51f468dd05f23c1db878fd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a77485fd9e246f190724b86663e0dbe81146e1fcc51f468dd05f23c1db878fd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adaa20d9a4898057802a8c9759ab3cd485963180bfc4f6b926bf6d2c886bcf27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adaa20d9a4898057802a8c9759ab3cd485963180bfc4f6b926bf6d2c886bcf27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://998fbb934f01c9ecd315f78e776199d4fd538106f8373f09b730d2df23ac9e3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://998fbb934f01c9ecd315f78e776199d4fd538106f8373f09b730d2df23ac9e3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46ebe18804ff07eb0e922a6cc1c39ff93875302d231222642b211e3527fe6373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46ebe18804ff07eb0e922a6cc1c39ff93875302d231222642b211e3527fe6373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfp2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:07Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:07 crc kubenswrapper[4948]: I0312 00:09:07.957082 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e483c2b-08f0-4e92-8e4a-b7281f30af3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d6839c99ab958ad1f40636c70e29bc285292fdbee1e910cdf7f8ab43ca7e8db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd9jv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82de8f2b633f1b6c295149d077bfd2b308f5acaa429acf411ed16f456536d0ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd9jv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4xwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:07Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:07 crc kubenswrapper[4948]: I0312 00:09:07.974496 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c835bb7a-4e06-4047-9fe6-e0c605b36bc6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc9760bd6ed0d092081fdd2ad33dda2d67f1305531640a35b9dce84b5d86a813\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f580535b16c31fb0ebadbdd9b7851ba2dce2f7d52bcad8c9cacc197f5e7dfad4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T00:07:08Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0312 00:06:43.492131 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0312 00:06:43.494442 1 observer_polling.go:159] Starting file observer\\\\nI0312 00:06:43.528056 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0312 00:06:43.536640 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0312 00:07:08.239474 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0312 00:07:08.239540 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb45482a404d6696cfbc3919997e53146b7b9dd1536056a55374574ae01ff193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://663bca87e44109c1d2bb59964e567d8b78fe2e9de39bcec0557cb43e39143add\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bf5b5d40a0cc76725ae2cf453089b31e75c304cf6037ef06b8ca8f7cff35a5a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:06:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:07Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:07 crc kubenswrapper[4948]: I0312 00:09:07.997020 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2efff53-8e12-41a1-bcb1-9a1d08c42a4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1cb328c8431b39dac735d49726a2eae40a367849b2c146bb2e49d02958f0e5a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ad13662849131222d5c8ba1041b81fa75bcb1fcbfa2e71c0c751f04b532a5b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8cd4faf47b25395ffd42d4bf5416ea29dcaa3ec380d6d6d6a9f13321d35195a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8028fd7b3e6d26bf30666425c843933b8fb87c3b2a258d510608ec718eb4fe82\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73834c8db23f6536034dce09e38b56665211238280bf9a4e1b13e74d81245a77\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T00:07:50Z\\\",\\\"message\\\":\\\"le observer\\\\nW0312 00:07:49.927227 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0312 00:07:49.927347 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 00:07:49.928000 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2511358960/tls.crt::/tmp/serving-cert-2511358960/tls.key\\\\\\\"\\\\nI0312 00:07:50.321455 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0312 00:07:50.324479 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0312 00:07:50.324511 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0312 00:07:50.324546 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0312 00:07:50.324556 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0312 00:07:50.331591 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0312 00:07:50.331609 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0312 00:07:50.331634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 00:07:50.331642 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 00:07:50.331651 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0312 00:07:50.331658 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0312 00:07:50.331664 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0312 00:07:50.331670 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0312 00:07:50.332917 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:07:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d56e32f9cb96ad103bab80d6a730b70d3387264a25a7a61ef1f0726e7290d6c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e42f2e94463ac8a3d7980a25122062da9ee87fcdc1f10e260ff9974b2e42359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e42f2e94463ac8a3d7980a25122062da9ee87fcdc1f10e260ff9974b2e42359\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:06:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:07Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:08 crc kubenswrapper[4948]: I0312 00:09:08.014773 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5a9ae9441886817563a5fc793bff7c9b8d77d406ceb1d1fb1eebd4307a1c64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:08Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:08 crc kubenswrapper[4948]: I0312 00:09:08.036189 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwpqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b7580c2-5c23-4c67-807a-ea97a3df9398\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b28502ee0add02df45488b29f20d8be4102b0f3faaf8b5e26b5943a410163bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d07c603619f858177d4b007e785208853abcfe2dc6c1af435309030fa409010\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T00:09:01Z\\\",\\\"message\\\":\\\"2026-03-12T00:08:16+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_5e485fc1-f819-4bac-9855-3dfbad154bba\\\\n2026-03-12T00:08:16+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_5e485fc1-f819-4bac-9855-3dfbad154bba to /host/opt/cni/bin/\\\\n2026-03-12T00:08:16Z [verbose] multus-daemon started\\\\n2026-03-12T00:08:16Z [verbose] Readiness Indicator file check\\\\n2026-03-12T00:09:01Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8lxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwpqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:08Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:08 crc kubenswrapper[4948]: I0312 00:09:08.054399 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-njc9c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0d6440c-4524-42ea-b18a-0ddc4662f672\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f4d0d9599ad62aaba35ba1414a806f3a163e29a3c34fd893ef5b14eae003591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7wn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-njc9c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:08Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:08 crc kubenswrapper[4948]: I0312 00:09:08.066082 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6c75836-fbdc-42f6-b581-7d1a827dc688\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f58f45d5eb63c29e243da4ed197ef81a00adb088ef9c8b65323dd3ff49f3c5a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78171d768ed45a0338b479044b9820500c88e17d0d2d9dcbf4b724de06ae3012\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78171d768ed45a0338b479044b9820500c88e17d0d2d9dcbf4b724de06ae3012\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:06:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:08Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:08 crc kubenswrapper[4948]: I0312 00:09:08.079731 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9s5fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f9c7f1-6593-4d92-854c-e6c9964a0435\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b8833cbaf3809463e18bfebc9d5bada5fb26823831cc742a3a27177662f60e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xf4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9s5fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:08Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:08 crc kubenswrapper[4948]: I0312 00:09:08.094615 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xdd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b5233d6-d2b8-42dc-b881-18e439ecc442\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0ea55d993c72b6d50a4829ecfff4adae7d0f625a90e40545c9e63201cdda29c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjpwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb23ff4bc484524e306561c0c574e2b2eee39fc80efcd9f44f1924fe42df1a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjpwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4xdd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:08Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:08 crc kubenswrapper[4948]: I0312 00:09:08.110084 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a124d35d-f697-49ae-8dac-ee32cd6fc5f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49c25293fe6ddb25b23fc250e3a5d94ce3f2c464f450c0da24dcff2febf15124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a57137a3fb48ab82c4324c6121b7dbcde8dd33e02cd7012ff42248cc5f164bec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccb319a44802931d007a0d9265f98bac6f23efee9a7fb82798718a315464e89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c756f1a5b429d17d95df4797725081682b66e9ef3e53a9deb9518e326e40325b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c756f1a5b429d17d95df4797725081682b66e9ef3e53a9deb9518e326e40325b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:06:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:08Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:08 crc kubenswrapper[4948]: I0312 00:09:08.126573 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:08Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:08 crc kubenswrapper[4948]: I0312 00:09:08.141406 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8d58ee45d75284e030700fa526b2c39197a26f6a391a7417f25a026f356b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:08Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:08 crc kubenswrapper[4948]: I0312 00:09:08.157400 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mqmtd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be985ff1-de8a-431f-a230-67894ebe52d5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmsz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmsz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mqmtd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:08Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:08 crc kubenswrapper[4948]: I0312 00:09:08.181684 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15117613-9fad-48c7-98c4-a2d84502ded9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4036d7131db00afd8eab8e093626b688297dc776b5af80c3c38b51da7124085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84af4cdd213d90636a889a8a8a3aaca3dee1e2677368efa0ffc768940c2cc994\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e481ace9e84b6aa66cbdda3ee8425c4120a665915d2c70ae4faf60bba6205e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9eadefd0babc119f776d48c9d2c9e4bced58db685af9a8e97f25e6a9a2e3c89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9d5e77753fda70350353daa0967257dd4929dba4440a95495ab4c2a0659aef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cad943edb9a8d445b33a6e70d85b475c1ccba460a6056bccae94fe60fd6b73f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d6a0a3fceb31bd31a6c978f0b3ac7ff768469ab3f311918e9dc72b424b00c27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a18330e68c4cbdf0380d4f4845bed125a80893c948f89c674632af8eb1ffb5f4\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T00:08:42Z\\\",\\\"message\\\":\\\"lector *v1.Namespace (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0312 00:08:42.392052 7170 reflector.go:311] Stopping reflector *v1.Node (0s) from k8s.io/client-go/informers/factory.go:160\\\\nI0312 00:08:42.392236 7170 reflector.go:311] Stopping reflector *v1.EgressService (0s) from github.com/openshift/ovn-kubernetes/go-controller/pkg/crd/egressservice/v1/apis/informers/externalversions/factory.go:140\\\\nI0312 00:08:42.392701 7170 handler.go:190] Sending *v1.EgressIP event handler 8 for removal\\\\nI0312 00:08:42.392763 7170 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0312 00:08:42.392775 7170 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0312 00:08:42.392819 7170 handler.go:190] Sending *v1.EgressFirewall event handler 9 for removal\\\\nI0312 00:08:42.392895 7170 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0312 00:08:42.392966 7170 factory.go:656] Stopping watch factory\\\\nI0312 00:08:42.392993 7170 ovnkube.go:599] Stopped ovnkube\\\\nI0312 00:08:42.393027 7170 handler.go:208] Removed *v1.EgressIP event handler 8\\\\nI0312 00:08:42.393049 7170 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0312 00:08:42.393062 7170 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0312 00:08:42.393074 7170 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0312 00:08:\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:41Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d6a0a3fceb31bd31a6c978f0b3ac7ff768469ab3f311918e9dc72b424b00c27\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T00:09:07Z\\\",\\\"message\\\":\\\"12 00:09:07.313053 7454 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0312 00:09:07.313076 7454 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0312 00:09:07.313090 7454 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0312 00:09:07.313112 7454 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0312 00:09:07.313116 7454 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0312 00:09:07.313125 7454 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0312 00:09:07.313112 7454 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0312 00:09:07.313125 7454 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0312 00:09:07.313145 7454 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0312 00:09:07.313199 7454 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0312 00:09:07.313208 7454 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0312 00:09:07.313223 7454 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0312 00:09:07.313256 7454 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0312 00:09:07.313259 7454 handler.go:208] Removed *v1.Node event handler 7\\\\nI0312 00:09:07.313337 7454 handler.go:208] Removed *v1.Node event handler 2\\\\nI0312 00:09:07.313507 7454 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:09:06Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46fa6140f4f4bd32e298fa36f6d7af6890029f9bc3080d2b4d94be8307346de8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7mp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:08Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:08 crc kubenswrapper[4948]: I0312 00:09:08.200968 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:08Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:08 crc kubenswrapper[4948]: I0312 00:09:08.311777 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:09:08 crc kubenswrapper[4948]: I0312 00:09:08.311810 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:09:08 crc kubenswrapper[4948]: I0312 00:09:08.311978 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:09:08 crc kubenswrapper[4948]: E0312 00:09:08.311978 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 00:09:08 crc kubenswrapper[4948]: I0312 00:09:08.312050 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:09:08 crc kubenswrapper[4948]: E0312 00:09:08.312282 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mqmtd" podUID="be985ff1-de8a-431f-a230-67894ebe52d5" Mar 12 00:09:08 crc kubenswrapper[4948]: E0312 00:09:08.312364 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 00:09:08 crc kubenswrapper[4948]: E0312 00:09:08.312472 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 00:09:08 crc kubenswrapper[4948]: I0312 00:09:08.885839 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f7mp5_15117613-9fad-48c7-98c4-a2d84502ded9/ovnkube-controller/3.log" Mar 12 00:09:08 crc kubenswrapper[4948]: I0312 00:09:08.890730 4948 scope.go:117] "RemoveContainer" containerID="1d6a0a3fceb31bd31a6c978f0b3ac7ff768469ab3f311918e9dc72b424b00c27" Mar 12 00:09:08 crc kubenswrapper[4948]: E0312 00:09:08.893395 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-f7mp5_openshift-ovn-kubernetes(15117613-9fad-48c7-98c4-a2d84502ded9)\"" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" podUID="15117613-9fad-48c7-98c4-a2d84502ded9" Mar 12 00:09:08 crc kubenswrapper[4948]: I0312 00:09:08.913924 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:08Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:08 crc kubenswrapper[4948]: I0312 00:09:08.936996 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfp2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"662dcbe7-7cf8-42d4-9a5e-345c32a49972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff3f3c69dcb28c7b0fd85f9e0c33b3339a1f6159dbf10bcf0efbdce4f7728c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a57d476b5b00110168878bd1c19c32ef122525c0259ac0a70de31c8220996d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4a57d476b5b00110168878bd1c19c32ef122525c0259ac0a70de31c8220996d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bb429b94a1286d8aac77cb77112c7163f7aa50a1c9cab997f8d088c100887c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0bb429b94a1286d8aac77cb77112c7163f7aa50a1c9cab997f8d088c100887c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a77485fd9e246f190724b86663e0dbe81146e1fcc51f468dd05f23c1db878fd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a77485fd9e246f190724b86663e0dbe81146e1fcc51f468dd05f23c1db878fd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adaa20d9a4898057802a8c9759ab3cd485963180bfc4f6b926bf6d2c886bcf27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adaa20d9a4898057802a8c9759ab3cd485963180bfc4f6b926bf6d2c886bcf27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://998fbb934f01c9ecd315f78e776199d4fd538106f8373f09b730d2df23ac9e3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://998fbb934f01c9ecd315f78e776199d4fd538106f8373f09b730d2df23ac9e3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46ebe18804ff07eb0e922a6cc1c39ff93875302d231222642b211e3527fe6373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46ebe18804ff07eb0e922a6cc1c39ff93875302d231222642b211e3527fe6373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfp2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:08Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:08 crc kubenswrapper[4948]: I0312 00:09:08.953920 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e483c2b-08f0-4e92-8e4a-b7281f30af3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d6839c99ab958ad1f40636c70e29bc285292fdbee1e910cdf7f8ab43ca7e8db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd9jv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82de8f2b633f1b6c295149d077bfd2b308f5acaa429acf411ed16f456536d0ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd9jv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4xwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:08Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:08 crc kubenswrapper[4948]: I0312 00:09:08.969986 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c835bb7a-4e06-4047-9fe6-e0c605b36bc6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc9760bd6ed0d092081fdd2ad33dda2d67f1305531640a35b9dce84b5d86a813\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f580535b16c31fb0ebadbdd9b7851ba2dce2f7d52bcad8c9cacc197f5e7dfad4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T00:07:08Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0312 00:06:43.492131 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0312 00:06:43.494442 1 observer_polling.go:159] Starting file observer\\\\nI0312 00:06:43.528056 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0312 00:06:43.536640 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0312 00:07:08.239474 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0312 00:07:08.239540 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb45482a404d6696cfbc3919997e53146b7b9dd1536056a55374574ae01ff193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://663bca87e44109c1d2bb59964e567d8b78fe2e9de39bcec0557cb43e39143add\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bf5b5d40a0cc76725ae2cf453089b31e75c304cf6037ef06b8ca8f7cff35a5a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:06:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:08Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:08 crc kubenswrapper[4948]: I0312 00:09:08.989555 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5139686878eb07d3abced92d0ccfe05be8907580ed75233b72fc2ad3415a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15cbe3fa75b69a7176803295e16a314a8bc73f5797c874124d54e76d47fa1064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:08Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:09 crc kubenswrapper[4948]: I0312 00:09:09.006095 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwpqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b7580c2-5c23-4c67-807a-ea97a3df9398\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b28502ee0add02df45488b29f20d8be4102b0f3faaf8b5e26b5943a410163bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d07c603619f858177d4b007e785208853abcfe2dc6c1af435309030fa409010\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T00:09:01Z\\\",\\\"message\\\":\\\"2026-03-12T00:08:16+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_5e485fc1-f819-4bac-9855-3dfbad154bba\\\\n2026-03-12T00:08:16+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_5e485fc1-f819-4bac-9855-3dfbad154bba to /host/opt/cni/bin/\\\\n2026-03-12T00:08:16Z [verbose] multus-daemon started\\\\n2026-03-12T00:08:16Z [verbose] Readiness Indicator file check\\\\n2026-03-12T00:09:01Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8lxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwpqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:09Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:09 crc kubenswrapper[4948]: I0312 00:09:09.020390 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-njc9c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0d6440c-4524-42ea-b18a-0ddc4662f672\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f4d0d9599ad62aaba35ba1414a806f3a163e29a3c34fd893ef5b14eae003591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7wn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-njc9c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:09Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:09 crc kubenswrapper[4948]: I0312 00:09:09.038627 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6c75836-fbdc-42f6-b581-7d1a827dc688\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f58f45d5eb63c29e243da4ed197ef81a00adb088ef9c8b65323dd3ff49f3c5a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78171d768ed45a0338b479044b9820500c88e17d0d2d9dcbf4b724de06ae3012\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78171d768ed45a0338b479044b9820500c88e17d0d2d9dcbf4b724de06ae3012\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:06:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:09Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:09 crc kubenswrapper[4948]: I0312 00:09:09.055836 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2efff53-8e12-41a1-bcb1-9a1d08c42a4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1cb328c8431b39dac735d49726a2eae40a367849b2c146bb2e49d02958f0e5a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ad13662849131222d5c8ba1041b81fa75bcb1fcbfa2e71c0c751f04b532a5b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8cd4faf47b25395ffd42d4bf5416ea29dcaa3ec380d6d6d6a9f13321d35195a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8028fd7b3e6d26bf30666425c843933b8fb87c3b2a258d510608ec718eb4fe82\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73834c8db23f6536034dce09e38b56665211238280bf9a4e1b13e74d81245a77\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T00:07:50Z\\\",\\\"message\\\":\\\"le observer\\\\nW0312 00:07:49.927227 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0312 00:07:49.927347 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 00:07:49.928000 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2511358960/tls.crt::/tmp/serving-cert-2511358960/tls.key\\\\\\\"\\\\nI0312 00:07:50.321455 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0312 00:07:50.324479 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0312 00:07:50.324511 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0312 00:07:50.324546 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0312 00:07:50.324556 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0312 00:07:50.331591 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0312 00:07:50.331609 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0312 00:07:50.331634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 00:07:50.331642 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 00:07:50.331651 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0312 00:07:50.331658 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0312 00:07:50.331664 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0312 00:07:50.331670 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0312 00:07:50.332917 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:07:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d56e32f9cb96ad103bab80d6a730b70d3387264a25a7a61ef1f0726e7290d6c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e42f2e94463ac8a3d7980a25122062da9ee87fcdc1f10e260ff9974b2e42359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e42f2e94463ac8a3d7980a25122062da9ee87fcdc1f10e260ff9974b2e42359\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:06:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:09Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:09 crc kubenswrapper[4948]: I0312 00:09:09.075128 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5a9ae9441886817563a5fc793bff7c9b8d77d406ceb1d1fb1eebd4307a1c64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:09Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:09 crc kubenswrapper[4948]: I0312 00:09:09.091910 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a124d35d-f697-49ae-8dac-ee32cd6fc5f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49c25293fe6ddb25b23fc250e3a5d94ce3f2c464f450c0da24dcff2febf15124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a57137a3fb48ab82c4324c6121b7dbcde8dd33e02cd7012ff42248cc5f164bec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccb319a44802931d007a0d9265f98bac6f23efee9a7fb82798718a315464e89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c756f1a5b429d17d95df4797725081682b66e9ef3e53a9deb9518e326e40325b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c756f1a5b429d17d95df4797725081682b66e9ef3e53a9deb9518e326e40325b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:06:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:09Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:09 crc kubenswrapper[4948]: I0312 00:09:09.106670 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9s5fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f9c7f1-6593-4d92-854c-e6c9964a0435\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b8833cbaf3809463e18bfebc9d5bada5fb26823831cc742a3a27177662f60e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xf4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9s5fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:09Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:09 crc kubenswrapper[4948]: I0312 00:09:09.121696 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xdd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b5233d6-d2b8-42dc-b881-18e439ecc442\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0ea55d993c72b6d50a4829ecfff4adae7d0f625a90e40545c9e63201cdda29c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjpwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb23ff4bc484524e306561c0c574e2b2eee39fc80efcd9f44f1924fe42df1a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjpwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4xdd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:09Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:09 crc kubenswrapper[4948]: I0312 00:09:09.140137 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mqmtd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be985ff1-de8a-431f-a230-67894ebe52d5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmsz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmsz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mqmtd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:09Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:09 crc kubenswrapper[4948]: I0312 00:09:09.169183 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15117613-9fad-48c7-98c4-a2d84502ded9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4036d7131db00afd8eab8e093626b688297dc776b5af80c3c38b51da7124085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84af4cdd213d90636a889a8a8a3aaca3dee1e2677368efa0ffc768940c2cc994\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e481ace9e84b6aa66cbdda3ee8425c4120a665915d2c70ae4faf60bba6205e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9eadefd0babc119f776d48c9d2c9e4bced58db685af9a8e97f25e6a9a2e3c89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9d5e77753fda70350353daa0967257dd4929dba4440a95495ab4c2a0659aef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cad943edb9a8d445b33a6e70d85b475c1ccba460a6056bccae94fe60fd6b73f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d6a0a3fceb31bd31a6c978f0b3ac7ff768469ab3f311918e9dc72b424b00c27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d6a0a3fceb31bd31a6c978f0b3ac7ff768469ab3f311918e9dc72b424b00c27\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T00:09:07Z\\\",\\\"message\\\":\\\"12 00:09:07.313053 7454 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0312 00:09:07.313076 7454 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0312 00:09:07.313090 7454 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0312 00:09:07.313112 7454 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0312 00:09:07.313116 7454 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0312 00:09:07.313125 7454 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0312 00:09:07.313112 7454 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0312 00:09:07.313125 7454 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0312 00:09:07.313145 7454 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0312 00:09:07.313199 7454 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0312 00:09:07.313208 7454 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0312 00:09:07.313223 7454 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0312 00:09:07.313256 7454 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0312 00:09:07.313259 7454 handler.go:208] Removed *v1.Node event handler 7\\\\nI0312 00:09:07.313337 7454 handler.go:208] Removed *v1.Node event handler 2\\\\nI0312 00:09:07.313507 7454 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:09:06Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-f7mp5_openshift-ovn-kubernetes(15117613-9fad-48c7-98c4-a2d84502ded9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46fa6140f4f4bd32e298fa36f6d7af6890029f9bc3080d2b4d94be8307346de8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7mp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:09Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:09 crc kubenswrapper[4948]: I0312 00:09:09.187202 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:09Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:09 crc kubenswrapper[4948]: I0312 00:09:09.204741 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:09Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:09 crc kubenswrapper[4948]: I0312 00:09:09.223115 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8d58ee45d75284e030700fa526b2c39197a26f6a391a7417f25a026f356b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:09Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:10 crc kubenswrapper[4948]: I0312 00:09:10.312530 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:09:10 crc kubenswrapper[4948]: I0312 00:09:10.312583 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:09:10 crc kubenswrapper[4948]: I0312 00:09:10.312555 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:09:10 crc kubenswrapper[4948]: I0312 00:09:10.312530 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:09:10 crc kubenswrapper[4948]: E0312 00:09:10.312717 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 00:09:10 crc kubenswrapper[4948]: E0312 00:09:10.312914 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mqmtd" podUID="be985ff1-de8a-431f-a230-67894ebe52d5" Mar 12 00:09:10 crc kubenswrapper[4948]: E0312 00:09:10.313069 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 00:09:10 crc kubenswrapper[4948]: E0312 00:09:10.313154 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 00:09:11 crc kubenswrapper[4948]: I0312 00:09:11.334775 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwpqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b7580c2-5c23-4c67-807a-ea97a3df9398\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b28502ee0add02df45488b29f20d8be4102b0f3faaf8b5e26b5943a410163bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d07c603619f858177d4b007e785208853abcfe2dc6c1af435309030fa409010\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T00:09:01Z\\\",\\\"message\\\":\\\"2026-03-12T00:08:16+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_5e485fc1-f819-4bac-9855-3dfbad154bba\\\\n2026-03-12T00:08:16+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_5e485fc1-f819-4bac-9855-3dfbad154bba to /host/opt/cni/bin/\\\\n2026-03-12T00:08:16Z [verbose] multus-daemon started\\\\n2026-03-12T00:08:16Z [verbose] Readiness Indicator file check\\\\n2026-03-12T00:09:01Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8lxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwpqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:11Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:11 crc kubenswrapper[4948]: I0312 00:09:11.352712 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-njc9c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0d6440c-4524-42ea-b18a-0ddc4662f672\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f4d0d9599ad62aaba35ba1414a806f3a163e29a3c34fd893ef5b14eae003591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7wn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-njc9c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:11Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:11 crc kubenswrapper[4948]: I0312 00:09:11.368562 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6c75836-fbdc-42f6-b581-7d1a827dc688\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f58f45d5eb63c29e243da4ed197ef81a00adb088ef9c8b65323dd3ff49f3c5a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78171d768ed45a0338b479044b9820500c88e17d0d2d9dcbf4b724de06ae3012\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78171d768ed45a0338b479044b9820500c88e17d0d2d9dcbf4b724de06ae3012\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:06:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:11Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:11 crc kubenswrapper[4948]: I0312 00:09:11.395909 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2efff53-8e12-41a1-bcb1-9a1d08c42a4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1cb328c8431b39dac735d49726a2eae40a367849b2c146bb2e49d02958f0e5a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ad13662849131222d5c8ba1041b81fa75bcb1fcbfa2e71c0c751f04b532a5b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8cd4faf47b25395ffd42d4bf5416ea29dcaa3ec380d6d6d6a9f13321d35195a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8028fd7b3e6d26bf30666425c843933b8fb87c3b2a258d510608ec718eb4fe82\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73834c8db23f6536034dce09e38b56665211238280bf9a4e1b13e74d81245a77\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T00:07:50Z\\\",\\\"message\\\":\\\"le observer\\\\nW0312 00:07:49.927227 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0312 00:07:49.927347 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 00:07:49.928000 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2511358960/tls.crt::/tmp/serving-cert-2511358960/tls.key\\\\\\\"\\\\nI0312 00:07:50.321455 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0312 00:07:50.324479 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0312 00:07:50.324511 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0312 00:07:50.324546 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0312 00:07:50.324556 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0312 00:07:50.331591 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0312 00:07:50.331609 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0312 00:07:50.331634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 00:07:50.331642 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 00:07:50.331651 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0312 00:07:50.331658 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0312 00:07:50.331664 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0312 00:07:50.331670 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0312 00:07:50.332917 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:07:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d56e32f9cb96ad103bab80d6a730b70d3387264a25a7a61ef1f0726e7290d6c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e42f2e94463ac8a3d7980a25122062da9ee87fcdc1f10e260ff9974b2e42359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e42f2e94463ac8a3d7980a25122062da9ee87fcdc1f10e260ff9974b2e42359\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:06:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:11Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:11 crc kubenswrapper[4948]: I0312 00:09:11.418486 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5a9ae9441886817563a5fc793bff7c9b8d77d406ceb1d1fb1eebd4307a1c64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:11Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:11 crc kubenswrapper[4948]: I0312 00:09:11.436739 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a124d35d-f697-49ae-8dac-ee32cd6fc5f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49c25293fe6ddb25b23fc250e3a5d94ce3f2c464f450c0da24dcff2febf15124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a57137a3fb48ab82c4324c6121b7dbcde8dd33e02cd7012ff42248cc5f164bec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccb319a44802931d007a0d9265f98bac6f23efee9a7fb82798718a315464e89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c756f1a5b429d17d95df4797725081682b66e9ef3e53a9deb9518e326e40325b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c756f1a5b429d17d95df4797725081682b66e9ef3e53a9deb9518e326e40325b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:06:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:11Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:11 crc kubenswrapper[4948]: E0312 00:09:11.443857 4948 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 12 00:09:11 crc kubenswrapper[4948]: I0312 00:09:11.452363 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9s5fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f9c7f1-6593-4d92-854c-e6c9964a0435\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b8833cbaf3809463e18bfebc9d5bada5fb26823831cc742a3a27177662f60e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xf4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9s5fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:11Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:11 crc kubenswrapper[4948]: I0312 00:09:11.470498 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xdd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b5233d6-d2b8-42dc-b881-18e439ecc442\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0ea55d993c72b6d50a4829ecfff4adae7d0f625a90e40545c9e63201cdda29c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjpwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb23ff4bc484524e306561c0c574e2b2eee39fc80efcd9f44f1924fe42df1a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjpwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4xdd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:11Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:11 crc kubenswrapper[4948]: I0312 00:09:11.485885 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mqmtd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be985ff1-de8a-431f-a230-67894ebe52d5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmsz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmsz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mqmtd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:11Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:11 crc kubenswrapper[4948]: I0312 00:09:11.516484 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15117613-9fad-48c7-98c4-a2d84502ded9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4036d7131db00afd8eab8e093626b688297dc776b5af80c3c38b51da7124085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84af4cdd213d90636a889a8a8a3aaca3dee1e2677368efa0ffc768940c2cc994\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e481ace9e84b6aa66cbdda3ee8425c4120a665915d2c70ae4faf60bba6205e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9eadefd0babc119f776d48c9d2c9e4bced58db685af9a8e97f25e6a9a2e3c89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9d5e77753fda70350353daa0967257dd4929dba4440a95495ab4c2a0659aef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cad943edb9a8d445b33a6e70d85b475c1ccba460a6056bccae94fe60fd6b73f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d6a0a3fceb31bd31a6c978f0b3ac7ff768469ab3f311918e9dc72b424b00c27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d6a0a3fceb31bd31a6c978f0b3ac7ff768469ab3f311918e9dc72b424b00c27\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T00:09:07Z\\\",\\\"message\\\":\\\"12 00:09:07.313053 7454 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0312 00:09:07.313076 7454 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0312 00:09:07.313090 7454 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0312 00:09:07.313112 7454 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0312 00:09:07.313116 7454 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0312 00:09:07.313125 7454 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0312 00:09:07.313112 7454 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0312 00:09:07.313125 7454 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0312 00:09:07.313145 7454 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0312 00:09:07.313199 7454 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0312 00:09:07.313208 7454 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0312 00:09:07.313223 7454 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0312 00:09:07.313256 7454 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0312 00:09:07.313259 7454 handler.go:208] Removed *v1.Node event handler 7\\\\nI0312 00:09:07.313337 7454 handler.go:208] Removed *v1.Node event handler 2\\\\nI0312 00:09:07.313507 7454 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:09:06Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-f7mp5_openshift-ovn-kubernetes(15117613-9fad-48c7-98c4-a2d84502ded9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46fa6140f4f4bd32e298fa36f6d7af6890029f9bc3080d2b4d94be8307346de8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7mp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:11Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:11 crc kubenswrapper[4948]: I0312 00:09:11.535571 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:11Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:11 crc kubenswrapper[4948]: I0312 00:09:11.550041 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:11Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:11 crc kubenswrapper[4948]: I0312 00:09:11.563978 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8d58ee45d75284e030700fa526b2c39197a26f6a391a7417f25a026f356b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:11Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:11 crc kubenswrapper[4948]: I0312 00:09:11.582110 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:11Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:11 crc kubenswrapper[4948]: I0312 00:09:11.603963 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfp2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"662dcbe7-7cf8-42d4-9a5e-345c32a49972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff3f3c69dcb28c7b0fd85f9e0c33b3339a1f6159dbf10bcf0efbdce4f7728c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a57d476b5b00110168878bd1c19c32ef122525c0259ac0a70de31c8220996d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4a57d476b5b00110168878bd1c19c32ef122525c0259ac0a70de31c8220996d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bb429b94a1286d8aac77cb77112c7163f7aa50a1c9cab997f8d088c100887c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0bb429b94a1286d8aac77cb77112c7163f7aa50a1c9cab997f8d088c100887c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a77485fd9e246f190724b86663e0dbe81146e1fcc51f468dd05f23c1db878fd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a77485fd9e246f190724b86663e0dbe81146e1fcc51f468dd05f23c1db878fd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adaa20d9a4898057802a8c9759ab3cd485963180bfc4f6b926bf6d2c886bcf27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adaa20d9a4898057802a8c9759ab3cd485963180bfc4f6b926bf6d2c886bcf27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://998fbb934f01c9ecd315f78e776199d4fd538106f8373f09b730d2df23ac9e3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://998fbb934f01c9ecd315f78e776199d4fd538106f8373f09b730d2df23ac9e3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46ebe18804ff07eb0e922a6cc1c39ff93875302d231222642b211e3527fe6373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46ebe18804ff07eb0e922a6cc1c39ff93875302d231222642b211e3527fe6373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfp2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:11Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:11 crc kubenswrapper[4948]: I0312 00:09:11.621341 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e483c2b-08f0-4e92-8e4a-b7281f30af3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d6839c99ab958ad1f40636c70e29bc285292fdbee1e910cdf7f8ab43ca7e8db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd9jv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82de8f2b633f1b6c295149d077bfd2b308f5acaa429acf411ed16f456536d0ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd9jv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4xwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:11Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:11 crc kubenswrapper[4948]: I0312 00:09:11.640279 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c835bb7a-4e06-4047-9fe6-e0c605b36bc6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc9760bd6ed0d092081fdd2ad33dda2d67f1305531640a35b9dce84b5d86a813\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f580535b16c31fb0ebadbdd9b7851ba2dce2f7d52bcad8c9cacc197f5e7dfad4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T00:07:08Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0312 00:06:43.492131 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0312 00:06:43.494442 1 observer_polling.go:159] Starting file observer\\\\nI0312 00:06:43.528056 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0312 00:06:43.536640 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0312 00:07:08.239474 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0312 00:07:08.239540 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb45482a404d6696cfbc3919997e53146b7b9dd1536056a55374574ae01ff193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://663bca87e44109c1d2bb59964e567d8b78fe2e9de39bcec0557cb43e39143add\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bf5b5d40a0cc76725ae2cf453089b31e75c304cf6037ef06b8ca8f7cff35a5a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:06:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:11Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:11 crc kubenswrapper[4948]: I0312 00:09:11.658556 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5139686878eb07d3abced92d0ccfe05be8907580ed75233b72fc2ad3415a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15cbe3fa75b69a7176803295e16a314a8bc73f5797c874124d54e76d47fa1064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:11Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:12 crc kubenswrapper[4948]: I0312 00:09:12.311982 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:09:12 crc kubenswrapper[4948]: I0312 00:09:12.312029 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:09:12 crc kubenswrapper[4948]: I0312 00:09:12.312092 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:09:12 crc kubenswrapper[4948]: I0312 00:09:12.312007 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:09:12 crc kubenswrapper[4948]: E0312 00:09:12.312187 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mqmtd" podUID="be985ff1-de8a-431f-a230-67894ebe52d5" Mar 12 00:09:12 crc kubenswrapper[4948]: E0312 00:09:12.312425 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 00:09:12 crc kubenswrapper[4948]: E0312 00:09:12.312568 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 00:09:12 crc kubenswrapper[4948]: E0312 00:09:12.312678 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 00:09:13 crc kubenswrapper[4948]: I0312 00:09:13.140144 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:09:13 crc kubenswrapper[4948]: I0312 00:09:13.140216 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:09:13 crc kubenswrapper[4948]: I0312 00:09:13.140241 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:09:13 crc kubenswrapper[4948]: I0312 00:09:13.140271 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:09:13 crc kubenswrapper[4948]: I0312 00:09:13.140292 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:09:13Z","lastTransitionTime":"2026-03-12T00:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:09:13 crc kubenswrapper[4948]: E0312 00:09:13.161029 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17825588-29cf-4059-bab7-1b042b9bf1f3\\\",\\\"systemUUID\\\":\\\"b18e4bf2-d37e-4b80-8da3-cd3803754ba2\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:13Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:13 crc kubenswrapper[4948]: I0312 00:09:13.166440 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:09:13 crc kubenswrapper[4948]: I0312 00:09:13.166517 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:09:13 crc kubenswrapper[4948]: I0312 00:09:13.166528 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:09:13 crc kubenswrapper[4948]: I0312 00:09:13.166545 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:09:13 crc kubenswrapper[4948]: I0312 00:09:13.166558 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:09:13Z","lastTransitionTime":"2026-03-12T00:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:09:13 crc kubenswrapper[4948]: E0312 00:09:13.186362 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17825588-29cf-4059-bab7-1b042b9bf1f3\\\",\\\"systemUUID\\\":\\\"b18e4bf2-d37e-4b80-8da3-cd3803754ba2\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:13Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:13 crc kubenswrapper[4948]: I0312 00:09:13.192697 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:09:13 crc kubenswrapper[4948]: I0312 00:09:13.192792 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:09:13 crc kubenswrapper[4948]: I0312 00:09:13.192818 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:09:13 crc kubenswrapper[4948]: I0312 00:09:13.192845 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:09:13 crc kubenswrapper[4948]: I0312 00:09:13.192866 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:09:13Z","lastTransitionTime":"2026-03-12T00:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:09:13 crc kubenswrapper[4948]: E0312 00:09:13.222728 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17825588-29cf-4059-bab7-1b042b9bf1f3\\\",\\\"systemUUID\\\":\\\"b18e4bf2-d37e-4b80-8da3-cd3803754ba2\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:13Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:13 crc kubenswrapper[4948]: I0312 00:09:13.227497 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:09:13 crc kubenswrapper[4948]: I0312 00:09:13.227569 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:09:13 crc kubenswrapper[4948]: I0312 00:09:13.227592 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:09:13 crc kubenswrapper[4948]: I0312 00:09:13.227622 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:09:13 crc kubenswrapper[4948]: I0312 00:09:13.227721 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:09:13Z","lastTransitionTime":"2026-03-12T00:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:09:13 crc kubenswrapper[4948]: E0312 00:09:13.249772 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17825588-29cf-4059-bab7-1b042b9bf1f3\\\",\\\"systemUUID\\\":\\\"b18e4bf2-d37e-4b80-8da3-cd3803754ba2\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:13Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:13 crc kubenswrapper[4948]: I0312 00:09:13.254664 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:09:13 crc kubenswrapper[4948]: I0312 00:09:13.254725 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:09:13 crc kubenswrapper[4948]: I0312 00:09:13.254750 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:09:13 crc kubenswrapper[4948]: I0312 00:09:13.254779 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:09:13 crc kubenswrapper[4948]: I0312 00:09:13.254801 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:09:13Z","lastTransitionTime":"2026-03-12T00:09:13Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:09:13 crc kubenswrapper[4948]: E0312 00:09:13.277941 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:13Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:13Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:09:13Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:13Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17825588-29cf-4059-bab7-1b042b9bf1f3\\\",\\\"systemUUID\\\":\\\"b18e4bf2-d37e-4b80-8da3-cd3803754ba2\\\"},\\\"runtimeHandlers\\\":[{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":false},\\\"name\\\":\\\"runc\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"crun\\\"},{\\\"features\\\":{\\\"recursiveReadOnlyMounts\\\":true,\\\"userNamespaces\\\":true},\\\"name\\\":\\\"\\\"}]}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:13Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:13 crc kubenswrapper[4948]: E0312 00:09:13.278171 4948 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 12 00:09:14 crc kubenswrapper[4948]: I0312 00:09:14.311826 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:09:14 crc kubenswrapper[4948]: I0312 00:09:14.311826 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:09:14 crc kubenswrapper[4948]: I0312 00:09:14.311838 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:09:14 crc kubenswrapper[4948]: I0312 00:09:14.311956 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:09:14 crc kubenswrapper[4948]: E0312 00:09:14.312133 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 00:09:14 crc kubenswrapper[4948]: E0312 00:09:14.312289 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 00:09:14 crc kubenswrapper[4948]: E0312 00:09:14.312669 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 00:09:14 crc kubenswrapper[4948]: E0312 00:09:14.312832 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mqmtd" podUID="be985ff1-de8a-431f-a230-67894ebe52d5" Mar 12 00:09:16 crc kubenswrapper[4948]: I0312 00:09:16.312084 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:09:16 crc kubenswrapper[4948]: I0312 00:09:16.312123 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:09:16 crc kubenswrapper[4948]: I0312 00:09:16.312177 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:09:16 crc kubenswrapper[4948]: I0312 00:09:16.312123 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:09:16 crc kubenswrapper[4948]: E0312 00:09:16.312380 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 00:09:16 crc kubenswrapper[4948]: E0312 00:09:16.312546 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mqmtd" podUID="be985ff1-de8a-431f-a230-67894ebe52d5" Mar 12 00:09:16 crc kubenswrapper[4948]: E0312 00:09:16.312780 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 00:09:16 crc kubenswrapper[4948]: E0312 00:09:16.312922 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 00:09:16 crc kubenswrapper[4948]: E0312 00:09:16.445851 4948 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 12 00:09:18 crc kubenswrapper[4948]: I0312 00:09:18.135756 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 00:09:18 crc kubenswrapper[4948]: E0312 00:09:18.136441 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 00:10:22.13589266 +0000 UTC m=+221.591496428 (durationBeforeRetry 1m4s). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:09:18 crc kubenswrapper[4948]: I0312 00:09:18.136539 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:09:18 crc kubenswrapper[4948]: I0312 00:09:18.136602 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:09:18 crc kubenswrapper[4948]: I0312 00:09:18.136664 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:09:18 crc kubenswrapper[4948]: I0312 00:09:18.136709 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:09:18 crc kubenswrapper[4948]: E0312 00:09:18.136731 4948 secret.go:188] Couldn't get secret openshift-network-console/networking-console-plugin-cert: object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 12 00:09:18 crc kubenswrapper[4948]: E0312 00:09:18.136813 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-12 00:10:22.136788457 +0000 UTC m=+221.592392225 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "networking-console-plugin-cert" (UniqueName: "kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin-cert" not registered Mar 12 00:09:18 crc kubenswrapper[4948]: E0312 00:09:18.136928 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 12 00:09:18 crc kubenswrapper[4948]: E0312 00:09:18.136954 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 12 00:09:18 crc kubenswrapper[4948]: E0312 00:09:18.136950 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/kube-root-ca.crt: object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered Mar 12 00:09:18 crc kubenswrapper[4948]: E0312 00:09:18.137003 4948 projected.go:288] Couldn't get configMap openshift-network-diagnostics/openshift-service-ca.crt: object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered Mar 12 00:09:18 crc kubenswrapper[4948]: E0312 00:09:18.137023 4948 projected.go:194] Error preparing data for projected volume kube-api-access-s2dwl for pod openshift-network-diagnostics/network-check-source-55646444c4-trplf: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 00:09:18 crc kubenswrapper[4948]: E0312 00:09:18.137021 4948 configmap.go:193] Couldn't get configMap openshift-network-console/networking-console-plugin: object "openshift-network-console"/"networking-console-plugin" not registered Mar 12 00:09:18 crc kubenswrapper[4948]: E0312 00:09:18.136973 4948 projected.go:194] Error preparing data for projected volume kube-api-access-cqllr for pod openshift-network-diagnostics/network-check-target-xd92c: [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 00:09:18 crc kubenswrapper[4948]: E0312 00:09:18.137110 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl podName:9d751cbb-f2e2-430d-9754-c882a5e924a5 nodeName:}" failed. No retries permitted until 2026-03-12 00:10:22.137078846 +0000 UTC m=+221.592682624 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-s2dwl" (UniqueName: "kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl") pod "network-check-source-55646444c4-trplf" (UID: "9d751cbb-f2e2-430d-9754-c882a5e924a5") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 00:09:18 crc kubenswrapper[4948]: E0312 00:09:18.137175 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf podName:5fe485a1-e14f-4c09-b5b9-f252bc42b7e8 nodeName:}" failed. No retries permitted until 2026-03-12 00:10:22.137137058 +0000 UTC m=+221.592740856 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "nginx-conf" (UniqueName: "kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf") pod "networking-console-plugin-85b44fc459-gdk6g" (UID: "5fe485a1-e14f-4c09-b5b9-f252bc42b7e8") : object "openshift-network-console"/"networking-console-plugin" not registered Mar 12 00:09:18 crc kubenswrapper[4948]: E0312 00:09:18.137234 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr podName:3b6479f0-333b-4a96-9adf-2099afdc2447 nodeName:}" failed. No retries permitted until 2026-03-12 00:10:22.137213071 +0000 UTC m=+221.592816979 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "kube-api-access-cqllr" (UniqueName: "kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr") pod "network-check-target-xd92c" (UID: "3b6479f0-333b-4a96-9adf-2099afdc2447") : [object "openshift-network-diagnostics"/"kube-root-ca.crt" not registered, object "openshift-network-diagnostics"/"openshift-service-ca.crt" not registered] Mar 12 00:09:18 crc kubenswrapper[4948]: I0312 00:09:18.237765 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/be985ff1-de8a-431f-a230-67894ebe52d5-metrics-certs\") pod \"network-metrics-daemon-mqmtd\" (UID: \"be985ff1-de8a-431f-a230-67894ebe52d5\") " pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:09:18 crc kubenswrapper[4948]: E0312 00:09:18.237977 4948 secret.go:188] Couldn't get secret openshift-multus/metrics-daemon-secret: object "openshift-multus"/"metrics-daemon-secret" not registered Mar 12 00:09:18 crc kubenswrapper[4948]: E0312 00:09:18.238121 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/secret/be985ff1-de8a-431f-a230-67894ebe52d5-metrics-certs podName:be985ff1-de8a-431f-a230-67894ebe52d5 nodeName:}" failed. No retries permitted until 2026-03-12 00:10:22.23808515 +0000 UTC m=+221.693688948 (durationBeforeRetry 1m4s). Error: MountVolume.SetUp failed for volume "metrics-certs" (UniqueName: "kubernetes.io/secret/be985ff1-de8a-431f-a230-67894ebe52d5-metrics-certs") pod "network-metrics-daemon-mqmtd" (UID: "be985ff1-de8a-431f-a230-67894ebe52d5") : object "openshift-multus"/"metrics-daemon-secret" not registered Mar 12 00:09:18 crc kubenswrapper[4948]: I0312 00:09:18.312130 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:09:18 crc kubenswrapper[4948]: I0312 00:09:18.312153 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:09:18 crc kubenswrapper[4948]: I0312 00:09:18.312202 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:09:18 crc kubenswrapper[4948]: I0312 00:09:18.312217 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:09:18 crc kubenswrapper[4948]: E0312 00:09:18.312447 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 00:09:18 crc kubenswrapper[4948]: E0312 00:09:18.312792 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mqmtd" podUID="be985ff1-de8a-431f-a230-67894ebe52d5" Mar 12 00:09:18 crc kubenswrapper[4948]: E0312 00:09:18.313006 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 00:09:18 crc kubenswrapper[4948]: E0312 00:09:18.313384 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 00:09:20 crc kubenswrapper[4948]: I0312 00:09:20.312517 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:09:20 crc kubenswrapper[4948]: I0312 00:09:20.312642 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:09:20 crc kubenswrapper[4948]: E0312 00:09:20.312702 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 00:09:20 crc kubenswrapper[4948]: I0312 00:09:20.312760 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:09:20 crc kubenswrapper[4948]: I0312 00:09:20.312756 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:09:20 crc kubenswrapper[4948]: E0312 00:09:20.312930 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 00:09:20 crc kubenswrapper[4948]: E0312 00:09:20.313035 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mqmtd" podUID="be985ff1-de8a-431f-a230-67894ebe52d5" Mar 12 00:09:20 crc kubenswrapper[4948]: E0312 00:09:20.313167 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 00:09:21 crc kubenswrapper[4948]: I0312 00:09:21.335555 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d2efff53-8e12-41a1-bcb1-9a1d08c42a4b\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:53Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1cb328c8431b39dac735d49726a2eae40a367849b2c146bb2e49d02958f0e5a2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"},{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://1ad13662849131222d5c8ba1041b81fa75bcb1fcbfa2e71c0c751f04b532a5b3\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-regeneration-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://c8cd4faf47b25395ffd42d4bf5416ea29dcaa3ec380d6d6d6a9f13321d35195a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://8028fd7b3e6d26bf30666425c843933b8fb87c3b2a258d510608ec718eb4fe82\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://73834c8db23f6536034dce09e38b56665211238280bf9a4e1b13e74d81245a77\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T00:07:50Z\\\",\\\"message\\\":\\\"le observer\\\\nW0312 00:07:49.927227 1 builder.go:272] unable to get owner reference (falling back to namespace): pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\nI0312 00:07:49.927347 1 builder.go:304] check-endpoints version 4.18.0-202502101302.p0.g763313c.assembly.stream.el9-763313c-763313c860ea43fcfc9b1ac00ebae096b57c078e\\\\nI0312 00:07:49.928000 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/tmp/serving-cert-2511358960/tls.crt::/tmp/serving-cert-2511358960/tls.key\\\\\\\"\\\\nI0312 00:07:50.321455 1 requestheader_controller.go:247] Loaded a new request header values for RequestHeaderAuthRequestController\\\\nI0312 00:07:50.324479 1 maxinflight.go:139] \\\\\\\"Initialized nonMutatingChan\\\\\\\" len=400\\\\nI0312 00:07:50.324511 1 maxinflight.go:145] \\\\\\\"Initialized mutatingChan\\\\\\\" len=200\\\\nI0312 00:07:50.324546 1 maxinflight.go:116] \\\\\\\"Set denominator for readonly requests\\\\\\\" limit=400\\\\nI0312 00:07:50.324556 1 maxinflight.go:120] \\\\\\\"Set denominator for mutating requests\\\\\\\" limit=200\\\\nI0312 00:07:50.331591 1 secure_serving.go:57] Forcing use of http/1.1 only\\\\nI0312 00:07:50.331609 1 genericapiserver.go:533] MuxAndDiscoveryComplete has all endpoints registered and discovery information is complete\\\\nW0312 00:07:50.331634 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_ECDSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 00:07:50.331642 1 secure_serving.go:69] Use of insecure cipher 'TLS_ECDHE_RSA_WITH_AES_128_CBC_SHA256' detected.\\\\nW0312 00:07:50.331651 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_GCM_SHA256' detected.\\\\nW0312 00:07:50.331658 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_GCM_SHA384' detected.\\\\nW0312 00:07:50.331664 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_128_CBC_SHA' detected.\\\\nW0312 00:07:50.331670 1 secure_serving.go:69] Use of insecure cipher 'TLS_RSA_WITH_AES_256_CBC_SHA' detected.\\\\nF0312 00:07:50.332917 1 cmd.go:182] pods \\\\\\\"kube-apiserver-crc\\\\\\\" not found\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:07:49Z\\\"}},\\\"name\\\":\\\"kube-apiserver-check-endpoints\\\",\\\"ready\\\":true,\\\"restartCount\\\":4,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:36Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://d56e32f9cb96ad103bab80d6a730b70d3387264a25a7a61ef1f0726e7290d6c1\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-apiserver-insecure-readyz\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:44Z\\\"}}}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://7e42f2e94463ac8a3d7980a25122062da9ee87fcdc1f10e260ff9974b2e42359\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://7e42f2e94463ac8a3d7980a25122062da9ee87fcdc1f10e260ff9974b2e42359\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/log/kube-apiserver\\\",\\\"name\\\":\\\"audit-dir\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:06:41Z\\\"}}\" for pod \"openshift-kube-apiserver\"/\"kube-apiserver-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:21Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:21 crc kubenswrapper[4948]: I0312 00:09:21.358704 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/network-operator-58b4c7f79c-55gtf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"37a5e44f-9a88-4405-be8a-b645485e7312\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://a5a9ae9441886817563a5fc793bff7c9b8d77d406ceb1d1fb1eebd4307a1c64b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-operator\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"host-etc-kube\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/serving-cert\\\",\\\"name\\\":\\\"metrics-tls\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rdwmf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"network-operator-58b4c7f79c-55gtf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:21Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:21 crc kubenswrapper[4948]: I0312 00:09:21.382601 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-bwpqp" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b7580c2-5c23-4c67-807a-ea97a3df9398\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:02Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://b28502ee0add02df45488b29f20d8be4102b0f3faaf8b5e26b5943a410163bf7\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://9d07c603619f858177d4b007e785208853abcfe2dc6c1af435309030fa409010\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T00:09:01Z\\\",\\\"message\\\":\\\"2026-03-12T00:08:16+00:00 [cnibincopy] Successfully copied files in /usr/src/multus-cni/rhel9/bin/ to /host/opt/cni/bin/upgrade_5e485fc1-f819-4bac-9855-3dfbad154bba\\\\n2026-03-12T00:08:16+00:00 [cnibincopy] Successfully moved files in /host/opt/cni/bin/upgrade_5e485fc1-f819-4bac-9855-3dfbad154bba to /host/opt/cni/bin/\\\\n2026-03-12T00:08:16Z [verbose] multus-daemon started\\\\n2026-03-12T00:08:16Z [verbose] Readiness Indicator file check\\\\n2026-03-12T00:09:01Z [error] have you checked that your default network is ready? still waiting for readinessindicatorfile @ /host/run/multus/cni/net.d/10-ovn-kubernetes.conf. pollimmediate error: timed out waiting for the condition\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"name\\\":\\\"kube-multus\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:09:01Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/run/multus/cni/net.d\\\",\\\"name\\\":\\\"multus-cni-dir\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/run/multus\\\",\\\"name\\\":\\\"multus-socket-dir-parent\\\"},{\\\"mountPath\\\":\\\"/run/k8s.cni.cncf.io\\\",\\\"name\\\":\\\"host-run-k8s-cni-cncf-io\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/bin\\\",\\\"name\\\":\\\"host-var-lib-cni-bin\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/multus\\\",\\\"name\\\":\\\"host-var-lib-cni-multus\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-var-lib-kubelet\\\"},{\\\"mountPath\\\":\\\"/hostroot\\\",\\\"name\\\":\\\"hostroot\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/net.d\\\",\\\"name\\\":\\\"multus-conf-dir\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d/multus.d\\\",\\\"name\\\":\\\"multus-daemon-config\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/cni/multus/certs\\\",\\\"name\\\":\\\"host-run-multus-certs\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kubernetes\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s8lxm\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-bwpqp\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:21Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:21 crc kubenswrapper[4948]: I0312 00:09:21.399175 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-image-registry/node-ca-njc9c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d0d6440c-4524-42ea-b18a-0ddc4662f672\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://6f4d0d9599ad62aaba35ba1414a806f3a163e29a3c34fd893ef5b14eae003591\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"node-ca\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/tmp/serviceca\\\",\\\"name\\\":\\\"serviceca\\\"},{\\\"mountPath\\\":\\\"/etc/docker/certs.d\\\",\\\"name\\\":\\\"host\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-k7wn2\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-image-registry\"/\"node-ca-njc9c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:21Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:21 crc kubenswrapper[4948]: I0312 00:09:21.417709 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"f6c75836-fbdc-42f6-b581-7d1a827dc688\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f58f45d5eb63c29e243da4ed197ef81a00adb088ef9c8b65323dd3ff49f3c5a0\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-crio\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes\\\",\\\"name\\\":\\\"etc-kube\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://78171d768ed45a0338b479044b9820500c88e17d0d2d9dcbf4b724de06ae3012\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://78171d768ed45a0338b479044b9820500c88e17d0d2d9dcbf4b724de06ae3012\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var\\\",\\\"name\\\":\\\"var-lib-kubelet\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:06:41Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"kube-rbac-proxy-crio-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:21Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:21 crc kubenswrapper[4948]: I0312 00:09:21.434893 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-dns/node-resolver-9s5fr" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"b9f9c7f1-6593-4d92-854c-e6c9964a0435\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://51b8833cbaf3809463e18bfebc9d5bada5fb26823831cc742a3a27177662f60e\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"dns-node-resolver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/hosts\\\",\\\"name\\\":\\\"hosts-file\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-9xf4z\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-dns\"/\"node-resolver-9s5fr\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:21Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:21 crc kubenswrapper[4948]: E0312 00:09:21.446690 4948 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 12 00:09:21 crc kubenswrapper[4948]: I0312 00:09:21.459885 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xdd8" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5b5233d6-d2b8-42dc-b881-18e439ecc442\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://f0ea55d993c72b6d50a4829ecfff4adae7d0f625a90e40545c9e63201cdda29c\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-control-plane-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjpwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://cb23ff4bc484524e306561c0c574e2b2eee39fc80efcd9f44f1924fe42df1a03\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovnkube-cluster-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hjpwd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-control-plane-749d76644c-4xdd8\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:21Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:21 crc kubenswrapper[4948]: I0312 00:09:21.479904 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"a124d35d-f697-49ae-8dac-ee32cd6fc5f1\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:43Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:31Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://49c25293fe6ddb25b23fc250e3a5d94ce3f2c464f450c0da24dcff2febf15124\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://a57137a3fb48ab82c4324c6121b7dbcde8dd33e02cd7012ff42248cc5f164bec\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://ccb319a44802931d007a0d9265f98bac6f23efee9a7fb82798718a315464e89f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-scheduler-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:44Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c756f1a5b429d17d95df4797725081682b66e9ef3e53a9deb9518e326e40325b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"wait-for-host-port\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c756f1a5b429d17d95df4797725081682b66e9ef3e53a9deb9518e326e40325b\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:06:42Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}}}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:06:41Z\\\"}}\" for pod \"openshift-kube-scheduler\"/\"openshift-kube-scheduler-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:21Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:21 crc kubenswrapper[4948]: I0312 00:09:21.500810 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"3b6479f0-333b-4a96-9adf-2099afdc2447\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-check-target-container]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"network-check-target-container\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-cqllr\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-target-xd92c\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:21Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:21 crc kubenswrapper[4948]: I0312 00:09:21.521242 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-operator/iptables-alerter-4ln5h" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"d75a4c96-2883-4a0b-bab2-0fab2b6c0b49\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://8c8d58ee45d75284e030700fa526b2c39197a26f6a391a7417f25a026f356b5b\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"iptables-alerter\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/iptables-alerter\\\",\\\"name\\\":\\\"iptables-alerter-script\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-rczfb\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-operator\"/\"iptables-alerter-4ln5h\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:21Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:21 crc kubenswrapper[4948]: I0312 00:09:21.536785 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/network-metrics-daemon-mqmtd" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"be985ff1-de8a-431f-a230-67894ebe52d5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [network-metrics-daemon kube-rbac-proxy]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/metrics\\\",\\\"name\\\":\\\"metrics-certs\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmsz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:d98bb346a17feae024d92663df92b25c120938395ab7043afbed543c6db9ca8d\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"network-metrics-daemon\\\",\\\"ready\\\":false,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-hmsz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"network-metrics-daemon-mqmtd\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:21Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:21 crc kubenswrapper[4948]: I0312 00:09:21.568106 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"15117613-9fad-48c7-98c4-a2d84502ded9\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [ovnkube-controller]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://d4036d7131db00afd8eab8e093626b688297dc776b5af80c3c38b51da7124085\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-node\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://84af4cdd213d90636a889a8a8a3aaca3dee1e2677368efa0ffc768940c2cc994\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy-ovn-metrics\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/pki/tls/metrics-cert\\\",\\\"name\\\":\\\"ovn-node-metrics-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://2e481ace9e84b6aa66cbdda3ee8425c4120a665915d2c70ae4faf60bba6205e2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"nbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://c9eadefd0babc119f776d48c9d2c9e4bced58db685af9a8e97f25e6a9a2e3c89\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"northd\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://ec9d5e77753fda70350353daa0967257dd4929dba4440a95495ab4c2a0659aef\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-acl-logging\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://8cad943edb9a8d445b33a6e70d85b475c1ccba460a6056bccae94fe60fd6b73f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"ovn-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn/\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/dev/log\\\",\\\"name\\\":\\\"log-socket\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://1d6a0a3fceb31bd31a6c978f0b3ac7ff768469ab3f311918e9dc72b424b00c27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://1d6a0a3fceb31bd31a6c978f0b3ac7ff768469ab3f311918e9dc72b424b00c27\\\",\\\"exitCode\\\":1,\\\"finishedAt\\\":\\\"2026-03-12T00:09:07Z\\\",\\\"message\\\":\\\"12 00:09:07.313053 7454 handler.go:190] Sending *v1.NetworkPolicy event handler 4 for removal\\\\nI0312 00:09:07.313076 7454 handler.go:190] Sending *v1.Namespace event handler 1 for removal\\\\nI0312 00:09:07.313090 7454 handler.go:190] Sending *v1.Namespace event handler 5 for removal\\\\nI0312 00:09:07.313112 7454 handler.go:190] Sending *v1.Node event handler 2 for removal\\\\nI0312 00:09:07.313116 7454 handler.go:208] Removed *v1.Namespace event handler 1\\\\nI0312 00:09:07.313125 7454 handler.go:208] Removed *v1.NetworkPolicy event handler 4\\\\nI0312 00:09:07.313112 7454 handler.go:208] Removed *v1.EgressFirewall event handler 9\\\\nI0312 00:09:07.313125 7454 handler.go:190] Sending *v1.Node event handler 7 for removal\\\\nI0312 00:09:07.313145 7454 handler.go:208] Removed *v1.Namespace event handler 5\\\\nI0312 00:09:07.313199 7454 handler.go:190] Sending *v1.Pod event handler 3 for removal\\\\nI0312 00:09:07.313208 7454 handler.go:190] Sending *v1.Pod event handler 6 for removal\\\\nI0312 00:09:07.313223 7454 handler.go:208] Removed *v1.Pod event handler 3\\\\nI0312 00:09:07.313256 7454 handler.go:208] Removed *v1.Pod event handler 6\\\\nI0312 00:09:07.313259 7454 handler.go:208] Removed *v1.Node event handler 7\\\\nI0312 00:09:07.313337 7454 handler.go:208] Removed *v1.Node event handler 2\\\\nI0312 00:09:07.313507 7454 factory.go:656] Stopping \\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:09:06Z\\\"}},\\\"name\\\":\\\"ovnkube-controller\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"message\\\":\\\"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-f7mp5_openshift-ovn-kubernetes(15117613-9fad-48c7-98c4-a2d84502ded9)\\\",\\\"reason\\\":\\\"CrashLoopBackOff\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/var/lib/kubelet\\\",\\\"name\\\":\\\"host-kubelet\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/systemd/system\\\",\\\"name\\\":\\\"systemd-units\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host\\\",\\\"name\\\":\\\"host-slash\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/ovn-kubernetes/\\\",\\\"name\\\":\\\"host-run-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/netns\\\",\\\"name\\\":\\\"host-run-netns\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/run/systemd/private\\\",\\\"name\\\":\\\"run-systemd\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/cni-bin-dir\\\",\\\"name\\\":\\\"host-cni-bin\\\"},{\\\"mountPath\\\":\\\"/etc/cni/net.d\\\",\\\"name\\\":\\\"host-cni-netd\\\"},{\\\"mountPath\\\":\\\"/var/lib/cni/networks/ovn-k8s-cni-overlay\\\",\\\"name\\\":\\\"host-var-lib-cni-networks-ovn-kubernetes\\\"},{\\\"mountPath\\\":\\\"/run/openvswitch\\\",\\\"name\\\":\\\"run-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/log/ovnkube/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/etc/openvswitch\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/lib/openvswitch\\\",\\\"name\\\":\\\"var-lib-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovnkube-config/\\\",\\\"name\\\":\\\"ovnkube-config\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46fa6140f4f4bd32e298fa36f6d7af6890029f9bc3080d2b4d94be8307346de8\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"sbdb\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/ovnkube-lib\\\",\\\"name\\\":\\\"ovnkube-script-lib\\\"},{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/run/ovn/\\\",\\\"name\\\":\\\"run-ovn\\\"},{\\\"mountPath\\\":\\\"/var/log/ovn\\\",\\\"name\\\":\\\"node-log\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kubecfg-setup\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/ovn/\\\",\\\"name\\\":\\\"etc-openvswitch\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-dvwhv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-ovn-kubernetes\"/\"ovnkube-node-f7mp5\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:21Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:21 crc kubenswrapper[4948]: I0312 00:09:21.591854 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"9d751cbb-f2e2-430d-9754-c882a5e924a5\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [check-endpoints]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"check-endpoints\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2dwl\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-diagnostics\"/\"network-check-source-55646444c4-trplf\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:21Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:21 crc kubenswrapper[4948]: I0312 00:09:21.614073 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-node-identity/network-node-identity-vrzqb" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"ef543e1b-8068-4ea3-b32a-61027b32e95d\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://1f5139686878eb07d3abced92d0ccfe05be8907580ed75233b72fc2ad3415a5a\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"approver\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://15cbe3fa75b69a7176803295e16a314a8bc73f5797c874124d54e76d47fa1064\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"webhook\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/webhook-cert/\\\",\\\"name\\\":\\\"webhook-cert\\\"},{\\\"mountPath\\\":\\\"/env\\\",\\\"name\\\":\\\"env-overrides\\\"},{\\\"mountPath\\\":\\\"/var/run/ovnkube-identity-config\\\",\\\"name\\\":\\\"ovnkube-identity-cm\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-s2kz5\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}]}}\" for pod \"openshift-network-node-identity\"/\"network-node-identity-vrzqb\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:21Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:21 crc kubenswrapper[4948]: I0312 00:09:21.638039 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"message\\\":\\\"containers with unready status: [networking-console-plugin]\\\",\\\"reason\\\":\\\"ContainersNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ae647598ec35cda5766806d3d44a91e3b9d4dee48ff154f3d8490165399873fd\\\",\\\"imageID\\\":\\\"\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"exitCode\\\":137,\\\"finishedAt\\\":null,\\\"message\\\":\\\"The container could not be located when the pod was deleted. The container used to be Running\\\",\\\"reason\\\":\\\"ContainerStatusUnknown\\\",\\\"startedAt\\\":null}},\\\"name\\\":\\\"networking-console-plugin\\\",\\\"ready\\\":false,\\\"restartCount\\\":3,\\\"started\\\":false,\\\"state\\\":{\\\"waiting\\\":{\\\"reason\\\":\\\"ContainerCreating\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/cert\\\",\\\"name\\\":\\\"networking-console-plugin-cert\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/etc/nginx/nginx.conf\\\",\\\"name\\\":\\\"nginx-conf\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"podIP\\\":null,\\\"podIPs\\\":null}}\" for pod \"openshift-network-console\"/\"networking-console-plugin-85b44fc459-gdk6g\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:21Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:21 crc kubenswrapper[4948]: I0312 00:09:21.664024 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-multus/multus-additional-cni-plugins-mfp2k" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"662dcbe7-7cf8-42d4-9a5e-345c32a49972\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:20Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://ff3f3c69dcb28c7b0fd85f9e0c33b3339a1f6159dbf10bcf0efbdce4f7728c86\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-multus-additional-cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:20Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"initContainerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://c4a57d476b5b00110168878bd1c19c32ef122525c0259ac0a70de31c8220996d\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"egress-router-binary-copy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://c4a57d476b5b00110168878bd1c19c32ef122525c0259ac0a70de31c8220996d\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bb429b94a1286d8aac77cb77112c7163f7aa50a1c9cab997f8d088c100887c2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"cni-plugins\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://0bb429b94a1286d8aac77cb77112c7163f7aa50a1c9cab997f8d088c100887c2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/tuning/\\\",\\\"name\\\":\\\"tuning-conf-dir\\\"},{\\\"mountPath\\\":\\\"/sysctls\\\",\\\"name\\\":\\\"cni-sysctl-allowlist\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://a77485fd9e246f190724b86663e0dbe81146e1fcc51f468dd05f23c1db878fd2\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:383f4cceeeaead203bb2327fdd367c64b64d729d7fa93089f249e496fcef0c78\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"bond-cni-plugin\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://a77485fd9e246f190724b86663e0dbe81146e1fcc51f468dd05f23c1db878fd2\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:16Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:16Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://adaa20d9a4898057802a8c9759ab3cd485963180bfc4f6b926bf6d2c886bcf27\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f567acb85146b5ed81451ec3e79f2de0c62e28c69b2eeade0abdf5d0c388e7aa\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"routeoverride-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://adaa20d9a4898057802a8c9759ab3cd485963180bfc4f6b926bf6d2c886bcf27\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:17Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:17Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://998fbb934f01c9ecd315f78e776199d4fd538106f8373f09b730d2df23ac9e3f\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni-bincopy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://998fbb934f01c9ecd315f78e776199d4fd538106f8373f09b730d2df23ac9e3f\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:18Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/entrypoint\\\",\\\"name\\\":\\\"cni-binary-copy\\\"},{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/os-release\\\",\\\"name\\\":\\\"os-release\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://46ebe18804ff07eb0e922a6cc1c39ff93875302d231222642b211e3527fe6373\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"whereabouts-cni\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":false,\\\"state\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://46ebe18804ff07eb0e922a6cc1c39ff93875302d231222642b211e3527fe6373\\\",\\\"exitCode\\\":0,\\\"finishedAt\\\":\\\"2026-03-12T00:08:19Z\\\",\\\"reason\\\":\\\"Completed\\\",\\\"startedAt\\\":\\\"2026-03-12T00:08:19Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/host/opt/cni/bin\\\",\\\"name\\\":\\\"cnibin\\\"},{\\\"mountPath\\\":\\\"/host/etc/cni/net.d\\\",\\\"name\\\":\\\"system-cni-dir\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-b6psg\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-multus\"/\"multus-additional-cni-plugins-mfp2k\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:21Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:21 crc kubenswrapper[4948]: I0312 00:09:21.685888 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"7e483c2b-08f0-4e92-8e4a-b7281f30af3e\\\"},\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"type\\\":\\\"Initialized\\\"},{\\\"type\\\":\\\"Ready\\\"},{\\\"type\\\":\\\"ContainersReady\\\"},{\\\"type\\\":\\\"PodScheduled\\\"}],\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:14Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:08:15Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://5d6839c99ab958ad1f40636c70e29bc285292fdbee1e910cdf7f8ab43ca7e8db\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-rbac-proxy\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:15Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/tls/private\\\",\\\"name\\\":\\\"proxy-tls\\\"},{\\\"mountPath\\\":\\\"/etc/kube-rbac-proxy\\\",\\\"name\\\":\\\"mcd-auth-proxy-config\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd9jv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]},{\\\"containerID\\\":\\\"cri-o://82de8f2b633f1b6c295149d077bfd2b308f5acaa429acf411ed16f456536d0ba\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"machine-config-daemon\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:08:14Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/rootfs\\\",\\\"name\\\":\\\"rootfs\\\"},{\\\"mountPath\\\":\\\"/var/run/secrets/kubernetes.io/serviceaccount\\\",\\\"name\\\":\\\"kube-api-access-gd9jv\\\",\\\"readOnly\\\":true,\\\"recursiveReadOnly\\\":\\\"Disabled\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:08:14Z\\\"}}\" for pod \"openshift-machine-config-operator\"/\"machine-config-daemon-m4xwc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:21Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:21 crc kubenswrapper[4948]: I0312 00:09:21.708447 4948 status_manager.go:875] "Failed to update status for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" err="failed to patch status \"{\\\"metadata\\\":{\\\"uid\\\":\\\"c835bb7a-4e06-4047-9fe6-e0c605b36bc6\\\"},\\\"status\\\":{\\\"conditions\\\":[{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:44Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodReadyToStartContainers\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Initialized\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"Ready\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:07:21Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"ContainersReady\\\"},{\\\"lastProbeTime\\\":null,\\\"lastTransitionTime\\\":\\\"2026-03-12T00:06:41Z\\\",\\\"status\\\":\\\"True\\\",\\\"type\\\":\\\"PodScheduled\\\"}],\\\"containerStatuses\\\":[{\\\"containerID\\\":\\\"cri-o://dc9760bd6ed0d092081fdd2ad33dda2d67f1305531640a35b9dce84b5d86a813\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\",\\\"lastState\\\":{\\\"terminated\\\":{\\\"containerID\\\":\\\"cri-o://f580535b16c31fb0ebadbdd9b7851ba2dce2f7d52bcad8c9cacc197f5e7dfad4\\\",\\\"exitCode\\\":255,\\\"finishedAt\\\":\\\"2026-03-12T00:07:08Z\\\",\\\"message\\\":\\\"+ timeout 3m /bin/bash -exuo pipefail -c 'while [ -n \\\\\\\"$(ss -Htanop \\\\\\\\( sport = 10357 \\\\\\\\))\\\\\\\" ]; do sleep 1; done'\\\\n++ ss -Htanop '(' sport = 10357 ')'\\\\n+ '[' -n '' ']'\\\\n+ exec cluster-policy-controller start --config=/etc/kubernetes/static-pod-resources/configmaps/cluster-policy-controller-config/config.yaml --kubeconfig=/etc/kubernetes/static-pod-resources/configmaps/controller-manager-kubeconfig/kubeconfig --namespace=openshift-kube-controller-manager -v=2\\\\nI0312 00:06:43.492131 1 leaderelection.go:121] The leader election gives 4 retries and allows for 30s of clock skew. The kube-apiserver downtime tolerance is 78s. Worst non-graceful lease acquisition is 2m43s. Worst graceful lease acquisition is {26s}.\\\\nI0312 00:06:43.494442 1 observer_polling.go:159] Starting file observer\\\\nI0312 00:06:43.528056 1 builder.go:298] cluster-policy-controller version 4.18.0-202501230001.p0.g5fd8525.assembly.stream.el9-5fd8525-5fd852525909ce6eab52972ba9ce8fcf56528eb9\\\\nI0312 00:06:43.536640 1 dynamic_serving_content.go:116] \\\\\\\"Loaded a new cert/key pair\\\\\\\" name=\\\\\\\"serving-cert::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.crt::/etc/kubernetes/static-pod-resources/secrets/serving-cert/tls.key\\\\\\\"\\\\nI0312 00:07:08.239474 1 cmd.go:138] Received SIGTERM or SIGINT signal, shutting down controller.\\\\nF0312 00:07:08.239540 1 cmd.go:179] failed checking apiserver connectivity: Unauthorized\\\\n\\\",\\\"reason\\\":\\\"Error\\\",\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}},\\\"name\\\":\\\"cluster-policy-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":1,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:07:10Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://fb45482a404d6696cfbc3919997e53146b7b9dd1536056a55374574ae01ff193\\\",\\\"image\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"imageID\\\":\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:42Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://663bca87e44109c1d2bb59964e567d8b78fe2e9de39bcec0557cb43e39143add\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-cert-syncer\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]},{\\\"containerID\\\":\\\"cri-o://0bf5b5d40a0cc76725ae2cf453089b31e75c304cf6037ef06b8ca8f7cff35a5a\\\",\\\"image\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"imageID\\\":\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\",\\\"lastState\\\":{},\\\"name\\\":\\\"kube-controller-manager-recovery-controller\\\",\\\"ready\\\":true,\\\"restartCount\\\":0,\\\"started\\\":true,\\\"state\\\":{\\\"running\\\":{\\\"startedAt\\\":\\\"2026-03-12T00:06:43Z\\\"}},\\\"volumeMounts\\\":[{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-resources\\\",\\\"name\\\":\\\"resource-dir\\\"},{\\\"mountPath\\\":\\\"/etc/kubernetes/static-pod-certs\\\",\\\"name\\\":\\\"cert-dir\\\"}]}],\\\"hostIP\\\":\\\"192.168.126.11\\\",\\\"hostIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"phase\\\":\\\"Running\\\",\\\"podIP\\\":\\\"192.168.126.11\\\",\\\"podIPs\\\":[{\\\"ip\\\":\\\"192.168.126.11\\\"}],\\\"startTime\\\":\\\"2026-03-12T00:06:41Z\\\"}}\" for pod \"openshift-kube-controller-manager\"/\"kube-controller-manager-crc\": Internal error occurred: failed calling webhook \"pod.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/pod?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:21Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:22 crc kubenswrapper[4948]: I0312 00:09:22.311884 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:09:22 crc kubenswrapper[4948]: I0312 00:09:22.311951 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:09:22 crc kubenswrapper[4948]: I0312 00:09:22.311922 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:09:22 crc kubenswrapper[4948]: E0312 00:09:22.312084 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 00:09:22 crc kubenswrapper[4948]: I0312 00:09:22.312132 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:09:22 crc kubenswrapper[4948]: E0312 00:09:22.312743 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 00:09:22 crc kubenswrapper[4948]: E0312 00:09:22.312851 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 00:09:22 crc kubenswrapper[4948]: E0312 00:09:22.312996 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mqmtd" podUID="be985ff1-de8a-431f-a230-67894ebe52d5" Mar 12 00:09:22 crc kubenswrapper[4948]: I0312 00:09:22.313228 4948 scope.go:117] "RemoveContainer" containerID="1d6a0a3fceb31bd31a6c978f0b3ac7ff768469ab3f311918e9dc72b424b00c27" Mar 12 00:09:22 crc kubenswrapper[4948]: E0312 00:09:22.313528 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-f7mp5_openshift-ovn-kubernetes(15117613-9fad-48c7-98c4-a2d84502ded9)\"" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" podUID="15117613-9fad-48c7-98c4-a2d84502ded9" Mar 12 00:09:23 crc kubenswrapper[4948]: I0312 00:09:23.436837 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:09:23 crc kubenswrapper[4948]: I0312 00:09:23.436915 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:09:23 crc kubenswrapper[4948]: I0312 00:09:23.436940 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:09:23 crc kubenswrapper[4948]: I0312 00:09:23.436969 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:09:23 crc kubenswrapper[4948]: I0312 00:09:23.436987 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:09:23Z","lastTransitionTime":"2026-03-12T00:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:09:23 crc kubenswrapper[4948]: E0312 00:09:23.458850 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17825588-29cf-4059-bab7-1b042b9bf1f3\\\",\\\"systemUUID\\\":\\\"b18e4bf2-d37e-4b80-8da3-cd3803754ba2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:23Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:23 crc kubenswrapper[4948]: I0312 00:09:23.464229 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:09:23 crc kubenswrapper[4948]: I0312 00:09:23.464294 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:09:23 crc kubenswrapper[4948]: I0312 00:09:23.464354 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:09:23 crc kubenswrapper[4948]: I0312 00:09:23.464386 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:09:23 crc kubenswrapper[4948]: I0312 00:09:23.464408 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:09:23Z","lastTransitionTime":"2026-03-12T00:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:09:23 crc kubenswrapper[4948]: E0312 00:09:23.489012 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17825588-29cf-4059-bab7-1b042b9bf1f3\\\",\\\"systemUUID\\\":\\\"b18e4bf2-d37e-4b80-8da3-cd3803754ba2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:23Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:23 crc kubenswrapper[4948]: I0312 00:09:23.495537 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:09:23 crc kubenswrapper[4948]: I0312 00:09:23.495765 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:09:23 crc kubenswrapper[4948]: I0312 00:09:23.495923 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:09:23 crc kubenswrapper[4948]: I0312 00:09:23.496079 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:09:23 crc kubenswrapper[4948]: I0312 00:09:23.496237 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:09:23Z","lastTransitionTime":"2026-03-12T00:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:09:23 crc kubenswrapper[4948]: E0312 00:09:23.516517 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17825588-29cf-4059-bab7-1b042b9bf1f3\\\",\\\"systemUUID\\\":\\\"b18e4bf2-d37e-4b80-8da3-cd3803754ba2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:23Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:23 crc kubenswrapper[4948]: I0312 00:09:23.520141 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:09:23 crc kubenswrapper[4948]: I0312 00:09:23.520179 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:09:23 crc kubenswrapper[4948]: I0312 00:09:23.520190 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:09:23 crc kubenswrapper[4948]: I0312 00:09:23.520207 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:09:23 crc kubenswrapper[4948]: I0312 00:09:23.520220 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:09:23Z","lastTransitionTime":"2026-03-12T00:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:09:23 crc kubenswrapper[4948]: E0312 00:09:23.535194 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17825588-29cf-4059-bab7-1b042b9bf1f3\\\",\\\"systemUUID\\\":\\\"b18e4bf2-d37e-4b80-8da3-cd3803754ba2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:23Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:23 crc kubenswrapper[4948]: I0312 00:09:23.539186 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:09:23 crc kubenswrapper[4948]: I0312 00:09:23.539242 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:09:23 crc kubenswrapper[4948]: I0312 00:09:23.539263 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:09:23 crc kubenswrapper[4948]: I0312 00:09:23.539288 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:09:23 crc kubenswrapper[4948]: I0312 00:09:23.539347 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:09:23Z","lastTransitionTime":"2026-03-12T00:09:23Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:09:23 crc kubenswrapper[4948]: E0312 00:09:23.560797 4948 kubelet_node_status.go:585] "Error updating node status, will retry" err="failed to patch status \"{\\\"status\\\":{\\\"$setElementOrder/conditions\\\":[{\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"type\\\":\\\"DiskPressure\\\"},{\\\"type\\\":\\\"PIDPressure\\\"},{\\\"type\\\":\\\"Ready\\\"}],\\\"allocatable\\\":{\\\"cpu\\\":\\\"11800m\\\",\\\"ephemeral-storage\\\":\\\"76396645454\\\",\\\"memory\\\":\\\"32404564Ki\\\"},\\\"capacity\\\":{\\\"cpu\\\":\\\"12\\\",\\\"ephemeral-storage\\\":\\\"83293888Ki\\\",\\\"memory\\\":\\\"32865364Ki\\\"},\\\"conditions\\\":[{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient memory available\\\",\\\"reason\\\":\\\"KubeletHasSufficientMemory\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"MemoryPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:23Z\\\",\\\"message\\\":\\\"kubelet has no disk pressure\\\",\\\"reason\\\":\\\"KubeletHasNoDiskPressure\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"DiskPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:23Z\\\",\\\"message\\\":\\\"kubelet has sufficient PID available\\\",\\\"reason\\\":\\\"KubeletHasSufficientPID\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"PIDPressure\\\"},{\\\"lastHeartbeatTime\\\":\\\"2026-03-12T00:09:23Z\\\",\\\"lastTransitionTime\\\":\\\"2026-03-12T00:09:23Z\\\",\\\"message\\\":\\\"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?\\\",\\\"reason\\\":\\\"KubeletNotReady\\\",\\\"status\\\":\\\"False\\\",\\\"type\\\":\\\"Ready\\\"}],\\\"images\\\":[{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b9ea248f8ca33258fe1683da51d2b16b94630be1b361c65f68a16c1a34b94887\\\"],\\\"sizeBytes\\\":2887430265},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:4a62fa1c0091f6d94e8fb7258470b9a532d78364b6b51a05341592041d598562\\\",\\\"registry.redhat.io/redhat/redhat-operator-index@sha256:8db792bab418e30d9b71b9e1ac330ad036025257abbd2cd32f318ed14f70d6ac\\\",\\\"registry.redhat.io/redhat/redhat-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1523204510},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:174f36cdd47ef0d1d2099482919d773257453265a2af0b17b154edc32fa41ac2\\\"],\\\"sizeBytes\\\":1498102846},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7eeaee65f2808b819eedb413bdcabb9144e12f0dd97f13fd1afba93a95b67b26\\\"],\\\"sizeBytes\\\":1232839934},{\\\"names\\\":[\\\"registry.redhat.io/redhat/community-operator-index@sha256:8ff55cdb2367f5011074d2f5ebdc153b8885e7495e14ae00f99d2b7ab3584ade\\\",\\\"registry.redhat.io/redhat/community-operator-index@sha256:d656c1453f2261d9b800f5c69fba3bc2ffdb388414c4c0e89fcbaa067d7614c4\\\",\\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1151049424},{\\\"names\\\":[\\\"registry.redhat.io/redhat/certified-operator-index@sha256:1d7d4739b2001bd173f2632d5f73724a5034237ee2d93a02a21bbfff547002ba\\\",\\\"registry.redhat.io/redhat/certified-operator-index@sha256:7688bce5eb0d153adff87fc9f7a47642465c0b88208efb236880197969931b37\\\",\\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"],\\\"sizeBytes\\\":1032059094},{\\\"names\\\":[\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:0878ac12c537fcfc617a539b3b8bd329ba568bb49c6e3bb47827b177c47ae669\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index@sha256:1dc15c170ebf462dacaef75511740ed94ca1da210f3980f66d77f91ba201c875\\\",\\\"registry.redhat.io/redhat/redhat-marketplace-index:v4.18\\\"],\\\"sizeBytes\\\":1001152198},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c915fb8ba96e911699a1ae34a8e95ca8a9fbe1bf8c28fea177225c63a8bdfc0a\\\"],\\\"sizeBytes\\\":964552795},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:06bc35825771aee1220d34720243b89c4ba8a8b335e6de2597126bd791fd90d4\\\"],\\\"sizeBytes\\\":947616130},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c3cc3840d7a81ce1b420f06e07a923861faf37d9c10688aa3aa0b7b76c8706ad\\\"],\\\"sizeBytes\\\":907837715},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:101f295e2eae0755ae1865f7de885db1f17b9368e4120a713bb5f79e17ce8f93\\\"],\\\"sizeBytes\\\":854694423},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:47b0670fa1051335fd2d2c9e8361e4ed77c7760c33a2180b136f7c7f59863ec2\\\"],\\\"sizeBytes\\\":852490370},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:862f4a4bed52f372056b6d368e2498ebfb063075b31cf48dbdaaeedfcf0396cb\\\"],\\\"sizeBytes\\\":772592048},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:98100674616e54319f6713d742fd0c3bdbc84e6e6173e8ccf4a2473a714c2bc4\\\"],\\\"sizeBytes\\\":705793115},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:687fddfbb085a1688df312ce4ec8c857df9b2daed8ff4a7ed6163a1154afa2cc\\\"],\\\"sizeBytes\\\":687915987},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:f247257b0885cf5d303e3612c7714b33ae51404cfa2429822060c6c025eb17dd\\\"],\\\"sizeBytes\\\":668060419},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e1baa38811c04bd8909e01a1f3be7421a1cb99d608d3dc4cf86d95b17de2ab8b\\\"],\\\"sizeBytes\\\":613826183},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e3e9dc0b02b9351edf7c46b1d46d724abd1ac38ecbd6bc541cee84a209258d8\\\"],\\\"sizeBytes\\\":581863411},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:35512335ac39aed0f55b7f799f416f4f6445c20c1b19888cf2bb72bb276703f2\\\"],\\\"sizeBytes\\\":574606365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:ee8d8f089ec1488067444c7e276c4e47cc93840280f3b3295484d67af2232002\\\"],\\\"sizeBytes\\\":550676059},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:10f20a39f16ae3019c62261eda8beb9e4d8c36cbb7b500b3bae1312987f0685d\\\"],\\\"sizeBytes\\\":541458174},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e40792096b162f0f9ce5f8362f51e5f8dea2c1ce4b1447235388416b5db7708c\\\"],\\\"sizeBytes\\\":533092226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:07b7c6877441ecd6a5646fb68e33e9be8b90092272e49117b54b4a67314731ca\\\"],\\\"sizeBytes\\\":528023732},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:a0fa3723269019bee1847b26702f42928e779036cc2f58408f8ee7866be30a93\\\"],\\\"sizeBytes\\\":510867594},{\\\"names\\\":[\\\"quay.io/crcont/ocp-release@sha256:0b6ae0d091d2bf49f9b3a3aff54aabdc49e70c783780f118789f49d8f95a9e03\\\"],\\\"sizeBytes\\\":510526836},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\\\"],\\\"sizeBytes\\\":507459597},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:7e9e7dd2b1a8394b7490ca6df8a3ee8cdfc6193ecc6fb6173ed9a1868116a207\\\"],\\\"sizeBytes\\\":505721947},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:094bb6a6641b4edbaf932f0551bcda20b0d4e012cbe84207348b24eeabd351e9\\\"],\\\"sizeBytes\\\":504778226},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c69fe7a98a744b7a7b61b2a8db81a338f373cd2b1d46c6d3f02864b30c37e46c\\\"],\\\"sizeBytes\\\":504735878},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:e51e6f78ec20ef91c82e94a49f950e427e77894e582dcc406eec4df807ddd76e\\\"],\\\"sizeBytes\\\":502943148},{\\\"names\\\":[\\\"quay.io/crcont/openshift-crc-cluster-kube-controller-manager-operator@sha256:8506ce0a578bc18fac117eb2b82799488ffac0bed08287faaf92edaf5d17ab95\\\"],\\\"sizeBytes\\\":501379880},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:3a741253807c962189819d879b8fef94a9452fb3f5f3969ec3207eb2d9862205\\\"],\\\"sizeBytes\\\":500472212},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5b881c97aa8e440c6b3ca001edfd789a9380066b8f11f35a8dd8d88c5c7dbf86\\\"],\\\"sizeBytes\\\":498888951},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:5aa9e5379bfeb63f4e517fb45168eb6820138041641bbdfc6f4db6427032fa37\\\"],\\\"sizeBytes\\\":497832828},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:c0f9da410c07372b6c9ad6a79379b491cd10fdee88051c026b084652d85aed21\\\"],\\\"sizeBytes\\\":497742284},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:88b1f0a05a1b1c91e1212b40f0e7d04c9351ec9d34c52097bfdc5897b46f2f0e\\\"],\\\"sizeBytes\\\":497120598},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:737e9019a072c74321e0a909ca95481f5c545044dd4f151a34d0e1c8b9cf273f\\\"],\\\"sizeBytes\\\":488494681},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:fe009d03910e18795e3bd60a3fd84938311d464d2730a2af5ded5b24e4d05a6b\\\"],\\\"sizeBytes\\\":487097366},{\\\"names\\\":[\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:66760a53b64d381940757ca9f0d05f523a61f943f8da03ce9791e5d05264a736\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner@sha256:e97a0cb5b6119a9735efe0ac24630a8912fcad89a1dddfa76dc10edac4ec9815\\\",\\\"registry.redhat.io/openshift4/ose-csi-external-provisioner:latest\\\"],\\\"sizeBytes\\\":485998616},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:9fa29d188c85a8b1e1bd15c9c18e96f1b235da9bd4a45dbc086a4a69520ed63f\\\"],\\\"sizeBytes\\\":485767738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:898cae57123c5006d397b24af21b0f24a0c42c9b0be5ee8251e1824711f65820\\\"],\\\"sizeBytes\\\":485535312},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:1eda5ad6a6c5b9cd94b4b456e9116f4a0517241b614de1a99df14baee20c3e6a\\\"],\\\"sizeBytes\\\":479585218},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:487c0a8d5200bcdce484ab1169229d8fcb8e91a934be45afff7819c4f7612f57\\\"],\\\"sizeBytes\\\":476681373},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:b641ed0d63034b23d07eb0b2cd455390e83b186e77375e2d3f37633c1ddb0495\\\"],\\\"sizeBytes\\\":473958144},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:32f9e10dfb8a7c812ea8b3e71a42bed9cef05305be18cc368b666df4643ba717\\\"],\\\"sizeBytes\\\":463179365},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:8fdf28927b06a42ea8af3985d558c84d9efd142bb32d3892c4fa9f5e0d98133c\\\"],\\\"sizeBytes\\\":460774792},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:dd0628f89ad843d82d5abfdc543ffab6a861a23cc3005909bd88fa7383b71113\\\"],\\\"sizeBytes\\\":459737917},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:242b3d66438c42745f4ef318bdeaf3d793426f12962a42ea83e18d06c08aaf09\\\"],\\\"sizeBytes\\\":457588564},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:adabc3456bf4f799f893d792cdf9e8cbc735b070be346552bcc99f741b0a83aa\\\"],\\\"sizeBytes\\\":450637738},{\\\"names\\\":[\\\"quay.io/openshift-release-dev/ocp-v4.0-art-dev@sha256:342dca43b5b09123737ccda5e41b4a5d564e54333d8ce04d867d3fb968600317\\\"],\\\"sizeBytes\\\":448887027}],\\\"nodeInfo\\\":{\\\"bootID\\\":\\\"17825588-29cf-4059-bab7-1b042b9bf1f3\\\",\\\"systemUUID\\\":\\\"b18e4bf2-d37e-4b80-8da3-cd3803754ba2\\\"}}}\" for node \"crc\": Internal error occurred: failed calling webhook \"node.network-node-identity.openshift.io\": failed to call webhook: Post \"https://127.0.0.1:9743/node?timeout=10s\": tls: failed to verify certificate: x509: certificate has expired or is not yet valid: current time 2026-03-12T00:09:23Z is after 2025-08-24T17:21:41Z" Mar 12 00:09:23 crc kubenswrapper[4948]: E0312 00:09:23.560941 4948 kubelet_node_status.go:572] "Unable to update node status" err="update node status exceeds retry count" Mar 12 00:09:24 crc kubenswrapper[4948]: I0312 00:09:24.311860 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:09:24 crc kubenswrapper[4948]: I0312 00:09:24.312007 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:09:24 crc kubenswrapper[4948]: I0312 00:09:24.312021 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:09:24 crc kubenswrapper[4948]: I0312 00:09:24.312030 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:09:24 crc kubenswrapper[4948]: E0312 00:09:24.312252 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 00:09:24 crc kubenswrapper[4948]: E0312 00:09:24.312507 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mqmtd" podUID="be985ff1-de8a-431f-a230-67894ebe52d5" Mar 12 00:09:24 crc kubenswrapper[4948]: E0312 00:09:24.313585 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 00:09:24 crc kubenswrapper[4948]: E0312 00:09:24.313748 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 00:09:26 crc kubenswrapper[4948]: I0312 00:09:26.312115 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:09:26 crc kubenswrapper[4948]: I0312 00:09:26.312168 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:09:26 crc kubenswrapper[4948]: E0312 00:09:26.313671 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mqmtd" podUID="be985ff1-de8a-431f-a230-67894ebe52d5" Mar 12 00:09:26 crc kubenswrapper[4948]: I0312 00:09:26.312218 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:09:26 crc kubenswrapper[4948]: E0312 00:09:26.313779 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 00:09:26 crc kubenswrapper[4948]: I0312 00:09:26.312166 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:09:26 crc kubenswrapper[4948]: E0312 00:09:26.313865 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 00:09:26 crc kubenswrapper[4948]: E0312 00:09:26.313993 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 00:09:26 crc kubenswrapper[4948]: E0312 00:09:26.448287 4948 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 12 00:09:28 crc kubenswrapper[4948]: I0312 00:09:28.312232 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:09:28 crc kubenswrapper[4948]: I0312 00:09:28.312271 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:09:28 crc kubenswrapper[4948]: I0312 00:09:28.312357 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:09:28 crc kubenswrapper[4948]: E0312 00:09:28.312445 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 00:09:28 crc kubenswrapper[4948]: I0312 00:09:28.312454 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:09:28 crc kubenswrapper[4948]: E0312 00:09:28.312673 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 00:09:28 crc kubenswrapper[4948]: E0312 00:09:28.313057 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 00:09:28 crc kubenswrapper[4948]: E0312 00:09:28.313180 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mqmtd" podUID="be985ff1-de8a-431f-a230-67894ebe52d5" Mar 12 00:09:28 crc kubenswrapper[4948]: I0312 00:09:28.332497 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd/etcd-crc"] Mar 12 00:09:30 crc kubenswrapper[4948]: I0312 00:09:30.312556 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:09:30 crc kubenswrapper[4948]: I0312 00:09:30.312634 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:09:30 crc kubenswrapper[4948]: I0312 00:09:30.312717 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:09:30 crc kubenswrapper[4948]: E0312 00:09:30.312815 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 00:09:30 crc kubenswrapper[4948]: I0312 00:09:30.312569 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:09:30 crc kubenswrapper[4948]: E0312 00:09:30.313025 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 00:09:30 crc kubenswrapper[4948]: E0312 00:09:30.313141 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 00:09:30 crc kubenswrapper[4948]: E0312 00:09:30.313284 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mqmtd" podUID="be985ff1-de8a-431f-a230-67894ebe52d5" Mar 12 00:09:31 crc kubenswrapper[4948]: I0312 00:09:31.382078 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd/etcd-crc" podStartSLOduration=3.382060084 podStartE2EDuration="3.382060084s" podCreationTimestamp="2026-03-12 00:09:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:09:31.381793916 +0000 UTC m=+170.837397654" watchObservedRunningTime="2026-03-12 00:09:31.382060084 +0000 UTC m=+170.837663822" Mar 12 00:09:31 crc kubenswrapper[4948]: I0312 00:09:31.382334 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podStartSLOduration=55.382327792 podStartE2EDuration="55.382327792s" podCreationTimestamp="2026-03-12 00:08:36 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:09:31.34357641 +0000 UTC m=+170.799180148" watchObservedRunningTime="2026-03-12 00:09:31.382327792 +0000 UTC m=+170.837931530" Mar 12 00:09:31 crc kubenswrapper[4948]: I0312 00:09:31.429077 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-additional-cni-plugins-mfp2k" podStartSLOduration=132.429055959 podStartE2EDuration="2m12.429055959s" podCreationTimestamp="2026-03-12 00:07:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:09:31.428028688 +0000 UTC m=+170.883632436" watchObservedRunningTime="2026-03-12 00:09:31.429055959 +0000 UTC m=+170.884659717" Mar 12 00:09:31 crc kubenswrapper[4948]: I0312 00:09:31.442062 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" podStartSLOduration=132.442043246 podStartE2EDuration="2m12.442043246s" podCreationTimestamp="2026-03-12 00:07:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:09:31.44183963 +0000 UTC m=+170.897443368" watchObservedRunningTime="2026-03-12 00:09:31.442043246 +0000 UTC m=+170.897646984" Mar 12 00:09:31 crc kubenswrapper[4948]: E0312 00:09:31.448990 4948 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 12 00:09:31 crc kubenswrapper[4948]: I0312 00:09:31.460217 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/kube-rbac-proxy-crio-crc" podStartSLOduration=33.46019827 podStartE2EDuration="33.46019827s" podCreationTimestamp="2026-03-12 00:08:58 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:09:31.460055906 +0000 UTC m=+170.915659674" watchObservedRunningTime="2026-03-12 00:09:31.46019827 +0000 UTC m=+170.915802018" Mar 12 00:09:31 crc kubenswrapper[4948]: I0312 00:09:31.474575 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=71.474555268 podStartE2EDuration="1m11.474555268s" podCreationTimestamp="2026-03-12 00:08:20 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:09:31.473797595 +0000 UTC m=+170.929401343" watchObservedRunningTime="2026-03-12 00:09:31.474555268 +0000 UTC m=+170.930159006" Mar 12 00:09:31 crc kubenswrapper[4948]: I0312 00:09:31.504214 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-bwpqp" podStartSLOduration=132.504190152 podStartE2EDuration="2m12.504190152s" podCreationTimestamp="2026-03-12 00:07:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:09:31.503700548 +0000 UTC m=+170.959304286" watchObservedRunningTime="2026-03-12 00:09:31.504190152 +0000 UTC m=+170.959793910" Mar 12 00:09:31 crc kubenswrapper[4948]: I0312 00:09:31.518636 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/node-ca-njc9c" podStartSLOduration=132.518621334 podStartE2EDuration="2m12.518621334s" podCreationTimestamp="2026-03-12 00:07:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:09:31.518063966 +0000 UTC m=+170.973667714" watchObservedRunningTime="2026-03-12 00:09:31.518621334 +0000 UTC m=+170.974225072" Mar 12 00:09:31 crc kubenswrapper[4948]: I0312 00:09:31.532525 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler/openshift-kube-scheduler-crc" podStartSLOduration=57.532508708 podStartE2EDuration="57.532508708s" podCreationTimestamp="2026-03-12 00:08:34 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:09:31.532023642 +0000 UTC m=+170.987627380" watchObservedRunningTime="2026-03-12 00:09:31.532508708 +0000 UTC m=+170.988112436" Mar 12 00:09:31 crc kubenswrapper[4948]: I0312 00:09:31.557920 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/node-resolver-9s5fr" podStartSLOduration=132.557898123 podStartE2EDuration="2m12.557898123s" podCreationTimestamp="2026-03-12 00:07:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:09:31.541614175 +0000 UTC m=+170.997217913" watchObservedRunningTime="2026-03-12 00:09:31.557898123 +0000 UTC m=+171.013501861" Mar 12 00:09:31 crc kubenswrapper[4948]: I0312 00:09:31.571513 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-control-plane-749d76644c-4xdd8" podStartSLOduration=132.571492677 podStartE2EDuration="2m12.571492677s" podCreationTimestamp="2026-03-12 00:07:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:09:31.559038668 +0000 UTC m=+171.014642406" watchObservedRunningTime="2026-03-12 00:09:31.571492677 +0000 UTC m=+171.027096425" Mar 12 00:09:32 crc kubenswrapper[4948]: I0312 00:09:32.311997 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:09:32 crc kubenswrapper[4948]: I0312 00:09:32.312012 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:09:32 crc kubenswrapper[4948]: I0312 00:09:32.312087 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:09:32 crc kubenswrapper[4948]: I0312 00:09:32.312094 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:09:32 crc kubenswrapper[4948]: E0312 00:09:32.312228 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 00:09:32 crc kubenswrapper[4948]: E0312 00:09:32.312671 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mqmtd" podUID="be985ff1-de8a-431f-a230-67894ebe52d5" Mar 12 00:09:32 crc kubenswrapper[4948]: E0312 00:09:32.312778 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 00:09:32 crc kubenswrapper[4948]: E0312 00:09:32.312795 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 00:09:33 crc kubenswrapper[4948]: I0312 00:09:33.312720 4948 scope.go:117] "RemoveContainer" containerID="1d6a0a3fceb31bd31a6c978f0b3ac7ff768469ab3f311918e9dc72b424b00c27" Mar 12 00:09:33 crc kubenswrapper[4948]: E0312 00:09:33.312976 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-f7mp5_openshift-ovn-kubernetes(15117613-9fad-48c7-98c4-a2d84502ded9)\"" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" podUID="15117613-9fad-48c7-98c4-a2d84502ded9" Mar 12 00:09:33 crc kubenswrapper[4948]: I0312 00:09:33.817024 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientMemory" Mar 12 00:09:33 crc kubenswrapper[4948]: I0312 00:09:33.817079 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasNoDiskPressure" Mar 12 00:09:33 crc kubenswrapper[4948]: I0312 00:09:33.817094 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeHasSufficientPID" Mar 12 00:09:33 crc kubenswrapper[4948]: I0312 00:09:33.817114 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeNotReady" Mar 12 00:09:33 crc kubenswrapper[4948]: I0312 00:09:33.817128 4948 setters.go:603] "Node became not ready" node="crc" condition={"type":"Ready","status":"False","lastHeartbeatTime":"2026-03-12T00:09:33Z","lastTransitionTime":"2026-03-12T00:09:33Z","reason":"KubeletNotReady","message":"container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?"} Mar 12 00:09:33 crc kubenswrapper[4948]: I0312 00:09:33.889796 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-version/cluster-version-operator-5c965bbfc6-wpchj"] Mar 12 00:09:33 crc kubenswrapper[4948]: I0312 00:09:33.890432 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wpchj" Mar 12 00:09:33 crc kubenswrapper[4948]: I0312 00:09:33.893853 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Mar 12 00:09:33 crc kubenswrapper[4948]: I0312 00:09:33.893889 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Mar 12 00:09:33 crc kubenswrapper[4948]: I0312 00:09:33.896429 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Mar 12 00:09:33 crc kubenswrapper[4948]: I0312 00:09:33.896553 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Mar 12 00:09:33 crc kubenswrapper[4948]: I0312 00:09:33.916826 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2035fcf4-f651-42f4-bdef-0cdea3e24a2d-service-ca\") pod \"cluster-version-operator-5c965bbfc6-wpchj\" (UID: \"2035fcf4-f651-42f4-bdef-0cdea3e24a2d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wpchj" Mar 12 00:09:33 crc kubenswrapper[4948]: I0312 00:09:33.916915 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/2035fcf4-f651-42f4-bdef-0cdea3e24a2d-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-wpchj\" (UID: \"2035fcf4-f651-42f4-bdef-0cdea3e24a2d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wpchj" Mar 12 00:09:33 crc kubenswrapper[4948]: I0312 00:09:33.916966 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2035fcf4-f651-42f4-bdef-0cdea3e24a2d-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-wpchj\" (UID: \"2035fcf4-f651-42f4-bdef-0cdea3e24a2d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wpchj" Mar 12 00:09:33 crc kubenswrapper[4948]: I0312 00:09:33.917020 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2035fcf4-f651-42f4-bdef-0cdea3e24a2d-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-wpchj\" (UID: \"2035fcf4-f651-42f4-bdef-0cdea3e24a2d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wpchj" Mar 12 00:09:33 crc kubenswrapper[4948]: I0312 00:09:33.917085 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/2035fcf4-f651-42f4-bdef-0cdea3e24a2d-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-wpchj\" (UID: \"2035fcf4-f651-42f4-bdef-0cdea3e24a2d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wpchj" Mar 12 00:09:34 crc kubenswrapper[4948]: I0312 00:09:34.017908 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2035fcf4-f651-42f4-bdef-0cdea3e24a2d-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-wpchj\" (UID: \"2035fcf4-f651-42f4-bdef-0cdea3e24a2d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wpchj" Mar 12 00:09:34 crc kubenswrapper[4948]: I0312 00:09:34.018006 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/2035fcf4-f651-42f4-bdef-0cdea3e24a2d-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-wpchj\" (UID: \"2035fcf4-f651-42f4-bdef-0cdea3e24a2d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wpchj" Mar 12 00:09:34 crc kubenswrapper[4948]: I0312 00:09:34.018048 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2035fcf4-f651-42f4-bdef-0cdea3e24a2d-service-ca\") pod \"cluster-version-operator-5c965bbfc6-wpchj\" (UID: \"2035fcf4-f651-42f4-bdef-0cdea3e24a2d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wpchj" Mar 12 00:09:34 crc kubenswrapper[4948]: I0312 00:09:34.018108 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/2035fcf4-f651-42f4-bdef-0cdea3e24a2d-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-wpchj\" (UID: \"2035fcf4-f651-42f4-bdef-0cdea3e24a2d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wpchj" Mar 12 00:09:34 crc kubenswrapper[4948]: I0312 00:09:34.018166 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2035fcf4-f651-42f4-bdef-0cdea3e24a2d-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-wpchj\" (UID: \"2035fcf4-f651-42f4-bdef-0cdea3e24a2d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wpchj" Mar 12 00:09:34 crc kubenswrapper[4948]: I0312 00:09:34.018215 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-ssl-certs\" (UniqueName: \"kubernetes.io/host-path/2035fcf4-f651-42f4-bdef-0cdea3e24a2d-etc-ssl-certs\") pod \"cluster-version-operator-5c965bbfc6-wpchj\" (UID: \"2035fcf4-f651-42f4-bdef-0cdea3e24a2d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wpchj" Mar 12 00:09:34 crc kubenswrapper[4948]: I0312 00:09:34.018367 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-cvo-updatepayloads\" (UniqueName: \"kubernetes.io/host-path/2035fcf4-f651-42f4-bdef-0cdea3e24a2d-etc-cvo-updatepayloads\") pod \"cluster-version-operator-5c965bbfc6-wpchj\" (UID: \"2035fcf4-f651-42f4-bdef-0cdea3e24a2d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wpchj" Mar 12 00:09:34 crc kubenswrapper[4948]: I0312 00:09:34.019078 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/2035fcf4-f651-42f4-bdef-0cdea3e24a2d-service-ca\") pod \"cluster-version-operator-5c965bbfc6-wpchj\" (UID: \"2035fcf4-f651-42f4-bdef-0cdea3e24a2d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wpchj" Mar 12 00:09:34 crc kubenswrapper[4948]: I0312 00:09:34.023557 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2035fcf4-f651-42f4-bdef-0cdea3e24a2d-serving-cert\") pod \"cluster-version-operator-5c965bbfc6-wpchj\" (UID: \"2035fcf4-f651-42f4-bdef-0cdea3e24a2d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wpchj" Mar 12 00:09:34 crc kubenswrapper[4948]: I0312 00:09:34.038273 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/2035fcf4-f651-42f4-bdef-0cdea3e24a2d-kube-api-access\") pod \"cluster-version-operator-5c965bbfc6-wpchj\" (UID: \"2035fcf4-f651-42f4-bdef-0cdea3e24a2d\") " pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wpchj" Mar 12 00:09:34 crc kubenswrapper[4948]: I0312 00:09:34.213233 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wpchj" Mar 12 00:09:34 crc kubenswrapper[4948]: W0312 00:09:34.226416 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2035fcf4_f651_42f4_bdef_0cdea3e24a2d.slice/crio-9d5f0cbc40c4b574edf6d0d386d52d874269da95a74c1411669ce563dd8b4552 WatchSource:0}: Error finding container 9d5f0cbc40c4b574edf6d0d386d52d874269da95a74c1411669ce563dd8b4552: Status 404 returned error can't find the container with id 9d5f0cbc40c4b574edf6d0d386d52d874269da95a74c1411669ce563dd8b4552 Mar 12 00:09:34 crc kubenswrapper[4948]: I0312 00:09:34.311710 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:09:34 crc kubenswrapper[4948]: E0312 00:09:34.312092 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 00:09:34 crc kubenswrapper[4948]: I0312 00:09:34.311816 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:09:34 crc kubenswrapper[4948]: E0312 00:09:34.312175 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 00:09:34 crc kubenswrapper[4948]: I0312 00:09:34.311776 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:09:34 crc kubenswrapper[4948]: I0312 00:09:34.311839 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:09:34 crc kubenswrapper[4948]: E0312 00:09:34.312269 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mqmtd" podUID="be985ff1-de8a-431f-a230-67894ebe52d5" Mar 12 00:09:34 crc kubenswrapper[4948]: E0312 00:09:34.312503 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 00:09:34 crc kubenswrapper[4948]: I0312 00:09:34.346379 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Rotating certificates Mar 12 00:09:34 crc kubenswrapper[4948]: I0312 00:09:34.354623 4948 reflector.go:368] Caches populated for *v1.CertificateSigningRequest from k8s.io/client-go/tools/watch/informerwatcher.go:146 Mar 12 00:09:34 crc kubenswrapper[4948]: I0312 00:09:34.989990 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wpchj" event={"ID":"2035fcf4-f651-42f4-bdef-0cdea3e24a2d","Type":"ContainerStarted","Data":"574aa8d52d1a1d206b28ddda1c471a0ba4923cb272145e48fe0fbfc3a040f649"} Mar 12 00:09:34 crc kubenswrapper[4948]: I0312 00:09:34.990090 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wpchj" event={"ID":"2035fcf4-f651-42f4-bdef-0cdea3e24a2d","Type":"ContainerStarted","Data":"9d5f0cbc40c4b574edf6d0d386d52d874269da95a74c1411669ce563dd8b4552"} Mar 12 00:09:35 crc kubenswrapper[4948]: I0312 00:09:35.013127 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-version/cluster-version-operator-5c965bbfc6-wpchj" podStartSLOduration=136.013100644 podStartE2EDuration="2m16.013100644s" podCreationTimestamp="2026-03-12 00:07:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:09:35.011987691 +0000 UTC m=+174.467591469" watchObservedRunningTime="2026-03-12 00:09:35.013100644 +0000 UTC m=+174.468704422" Mar 12 00:09:36 crc kubenswrapper[4948]: I0312 00:09:36.311496 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:09:36 crc kubenswrapper[4948]: I0312 00:09:36.311633 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:09:36 crc kubenswrapper[4948]: I0312 00:09:36.311824 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:09:36 crc kubenswrapper[4948]: I0312 00:09:36.311843 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:09:36 crc kubenswrapper[4948]: E0312 00:09:36.311799 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mqmtd" podUID="be985ff1-de8a-431f-a230-67894ebe52d5" Mar 12 00:09:36 crc kubenswrapper[4948]: E0312 00:09:36.312043 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 00:09:36 crc kubenswrapper[4948]: E0312 00:09:36.312147 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 00:09:36 crc kubenswrapper[4948]: E0312 00:09:36.312211 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 00:09:36 crc kubenswrapper[4948]: E0312 00:09:36.451496 4948 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 12 00:09:38 crc kubenswrapper[4948]: I0312 00:09:38.311519 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:09:38 crc kubenswrapper[4948]: I0312 00:09:38.311558 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:09:38 crc kubenswrapper[4948]: I0312 00:09:38.311595 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:09:38 crc kubenswrapper[4948]: I0312 00:09:38.311673 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:09:38 crc kubenswrapper[4948]: E0312 00:09:38.311872 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mqmtd" podUID="be985ff1-de8a-431f-a230-67894ebe52d5" Mar 12 00:09:38 crc kubenswrapper[4948]: E0312 00:09:38.312070 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 00:09:38 crc kubenswrapper[4948]: E0312 00:09:38.312192 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 00:09:38 crc kubenswrapper[4948]: E0312 00:09:38.312361 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 00:09:40 crc kubenswrapper[4948]: I0312 00:09:40.312057 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:09:40 crc kubenswrapper[4948]: I0312 00:09:40.312180 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:09:40 crc kubenswrapper[4948]: E0312 00:09:40.312261 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 00:09:40 crc kubenswrapper[4948]: I0312 00:09:40.312278 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:09:40 crc kubenswrapper[4948]: I0312 00:09:40.312334 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:09:40 crc kubenswrapper[4948]: E0312 00:09:40.312425 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 00:09:40 crc kubenswrapper[4948]: E0312 00:09:40.312690 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mqmtd" podUID="be985ff1-de8a-431f-a230-67894ebe52d5" Mar 12 00:09:40 crc kubenswrapper[4948]: E0312 00:09:40.312855 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 00:09:41 crc kubenswrapper[4948]: E0312 00:09:41.452277 4948 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 12 00:09:42 crc kubenswrapper[4948]: I0312 00:09:42.311515 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:09:42 crc kubenswrapper[4948]: I0312 00:09:42.311580 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:09:42 crc kubenswrapper[4948]: E0312 00:09:42.312228 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 00:09:42 crc kubenswrapper[4948]: I0312 00:09:42.311621 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:09:42 crc kubenswrapper[4948]: I0312 00:09:42.311594 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:09:42 crc kubenswrapper[4948]: E0312 00:09:42.312848 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 00:09:42 crc kubenswrapper[4948]: E0312 00:09:42.312872 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mqmtd" podUID="be985ff1-de8a-431f-a230-67894ebe52d5" Mar 12 00:09:42 crc kubenswrapper[4948]: E0312 00:09:42.312279 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 00:09:44 crc kubenswrapper[4948]: I0312 00:09:44.311662 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:09:44 crc kubenswrapper[4948]: I0312 00:09:44.311776 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:09:44 crc kubenswrapper[4948]: I0312 00:09:44.311776 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:09:44 crc kubenswrapper[4948]: I0312 00:09:44.311886 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:09:44 crc kubenswrapper[4948]: E0312 00:09:44.311878 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 00:09:44 crc kubenswrapper[4948]: E0312 00:09:44.312017 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mqmtd" podUID="be985ff1-de8a-431f-a230-67894ebe52d5" Mar 12 00:09:44 crc kubenswrapper[4948]: E0312 00:09:44.312148 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 00:09:44 crc kubenswrapper[4948]: E0312 00:09:44.312703 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 00:09:46 crc kubenswrapper[4948]: I0312 00:09:46.312513 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:09:46 crc kubenswrapper[4948]: I0312 00:09:46.312632 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:09:46 crc kubenswrapper[4948]: E0312 00:09:46.312810 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 00:09:46 crc kubenswrapper[4948]: I0312 00:09:46.313106 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:09:46 crc kubenswrapper[4948]: I0312 00:09:46.313146 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:09:46 crc kubenswrapper[4948]: E0312 00:09:46.313373 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mqmtd" podUID="be985ff1-de8a-431f-a230-67894ebe52d5" Mar 12 00:09:46 crc kubenswrapper[4948]: E0312 00:09:46.313684 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 00:09:46 crc kubenswrapper[4948]: E0312 00:09:46.314538 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 00:09:46 crc kubenswrapper[4948]: I0312 00:09:46.314955 4948 scope.go:117] "RemoveContainer" containerID="1d6a0a3fceb31bd31a6c978f0b3ac7ff768469ab3f311918e9dc72b424b00c27" Mar 12 00:09:46 crc kubenswrapper[4948]: E0312 00:09:46.315286 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"ovnkube-controller\" with CrashLoopBackOff: \"back-off 40s restarting failed container=ovnkube-controller pod=ovnkube-node-f7mp5_openshift-ovn-kubernetes(15117613-9fad-48c7-98c4-a2d84502ded9)\"" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" podUID="15117613-9fad-48c7-98c4-a2d84502ded9" Mar 12 00:09:46 crc kubenswrapper[4948]: E0312 00:09:46.453254 4948 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 12 00:09:48 crc kubenswrapper[4948]: I0312 00:09:48.039482 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-bwpqp_3b7580c2-5c23-4c67-807a-ea97a3df9398/kube-multus/1.log" Mar 12 00:09:48 crc kubenswrapper[4948]: I0312 00:09:48.040056 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-bwpqp_3b7580c2-5c23-4c67-807a-ea97a3df9398/kube-multus/0.log" Mar 12 00:09:48 crc kubenswrapper[4948]: I0312 00:09:48.040096 4948 generic.go:334] "Generic (PLEG): container finished" podID="3b7580c2-5c23-4c67-807a-ea97a3df9398" containerID="b28502ee0add02df45488b29f20d8be4102b0f3faaf8b5e26b5943a410163bf7" exitCode=1 Mar 12 00:09:48 crc kubenswrapper[4948]: I0312 00:09:48.040131 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-bwpqp" event={"ID":"3b7580c2-5c23-4c67-807a-ea97a3df9398","Type":"ContainerDied","Data":"b28502ee0add02df45488b29f20d8be4102b0f3faaf8b5e26b5943a410163bf7"} Mar 12 00:09:48 crc kubenswrapper[4948]: I0312 00:09:48.040206 4948 scope.go:117] "RemoveContainer" containerID="9d07c603619f858177d4b007e785208853abcfe2dc6c1af435309030fa409010" Mar 12 00:09:48 crc kubenswrapper[4948]: I0312 00:09:48.040570 4948 scope.go:117] "RemoveContainer" containerID="b28502ee0add02df45488b29f20d8be4102b0f3faaf8b5e26b5943a410163bf7" Mar 12 00:09:48 crc kubenswrapper[4948]: E0312 00:09:48.040726 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 10s restarting failed container=kube-multus pod=multus-bwpqp_openshift-multus(3b7580c2-5c23-4c67-807a-ea97a3df9398)\"" pod="openshift-multus/multus-bwpqp" podUID="3b7580c2-5c23-4c67-807a-ea97a3df9398" Mar 12 00:09:48 crc kubenswrapper[4948]: I0312 00:09:48.311745 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:09:48 crc kubenswrapper[4948]: I0312 00:09:48.311772 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:09:48 crc kubenswrapper[4948]: E0312 00:09:48.312249 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 00:09:48 crc kubenswrapper[4948]: I0312 00:09:48.311873 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:09:48 crc kubenswrapper[4948]: I0312 00:09:48.311804 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:09:48 crc kubenswrapper[4948]: E0312 00:09:48.312427 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mqmtd" podUID="be985ff1-de8a-431f-a230-67894ebe52d5" Mar 12 00:09:48 crc kubenswrapper[4948]: E0312 00:09:48.312560 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 00:09:48 crc kubenswrapper[4948]: E0312 00:09:48.312680 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 00:09:49 crc kubenswrapper[4948]: I0312 00:09:49.046290 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-bwpqp_3b7580c2-5c23-4c67-807a-ea97a3df9398/kube-multus/1.log" Mar 12 00:09:50 crc kubenswrapper[4948]: I0312 00:09:50.311587 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:09:50 crc kubenswrapper[4948]: I0312 00:09:50.311652 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:09:50 crc kubenswrapper[4948]: I0312 00:09:50.311619 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:09:50 crc kubenswrapper[4948]: I0312 00:09:50.311740 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:09:50 crc kubenswrapper[4948]: E0312 00:09:50.312192 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 00:09:50 crc kubenswrapper[4948]: E0312 00:09:50.312375 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mqmtd" podUID="be985ff1-de8a-431f-a230-67894ebe52d5" Mar 12 00:09:50 crc kubenswrapper[4948]: E0312 00:09:50.312549 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 00:09:50 crc kubenswrapper[4948]: E0312 00:09:50.312704 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 00:09:51 crc kubenswrapper[4948]: E0312 00:09:51.454188 4948 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 12 00:09:52 crc kubenswrapper[4948]: I0312 00:09:52.311632 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:09:52 crc kubenswrapper[4948]: I0312 00:09:52.311646 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:09:52 crc kubenswrapper[4948]: I0312 00:09:52.311782 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:09:52 crc kubenswrapper[4948]: I0312 00:09:52.311673 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:09:52 crc kubenswrapper[4948]: E0312 00:09:52.311855 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 00:09:52 crc kubenswrapper[4948]: E0312 00:09:52.311998 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 00:09:52 crc kubenswrapper[4948]: E0312 00:09:52.312110 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mqmtd" podUID="be985ff1-de8a-431f-a230-67894ebe52d5" Mar 12 00:09:52 crc kubenswrapper[4948]: E0312 00:09:52.312237 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 00:09:54 crc kubenswrapper[4948]: I0312 00:09:54.312425 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:09:54 crc kubenswrapper[4948]: I0312 00:09:54.312551 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:09:54 crc kubenswrapper[4948]: I0312 00:09:54.312452 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:09:54 crc kubenswrapper[4948]: I0312 00:09:54.312661 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:09:54 crc kubenswrapper[4948]: E0312 00:09:54.312624 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 00:09:54 crc kubenswrapper[4948]: E0312 00:09:54.312856 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 00:09:54 crc kubenswrapper[4948]: E0312 00:09:54.312917 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 00:09:54 crc kubenswrapper[4948]: E0312 00:09:54.313183 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mqmtd" podUID="be985ff1-de8a-431f-a230-67894ebe52d5" Mar 12 00:09:56 crc kubenswrapper[4948]: I0312 00:09:56.312044 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:09:56 crc kubenswrapper[4948]: I0312 00:09:56.312095 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:09:56 crc kubenswrapper[4948]: I0312 00:09:56.312124 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:09:56 crc kubenswrapper[4948]: I0312 00:09:56.312050 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:09:56 crc kubenswrapper[4948]: E0312 00:09:56.312237 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 00:09:56 crc kubenswrapper[4948]: E0312 00:09:56.312510 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 00:09:56 crc kubenswrapper[4948]: E0312 00:09:56.312643 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mqmtd" podUID="be985ff1-de8a-431f-a230-67894ebe52d5" Mar 12 00:09:56 crc kubenswrapper[4948]: E0312 00:09:56.312782 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 00:09:56 crc kubenswrapper[4948]: E0312 00:09:56.456500 4948 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 12 00:09:58 crc kubenswrapper[4948]: I0312 00:09:58.311668 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:09:58 crc kubenswrapper[4948]: I0312 00:09:58.311722 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:09:58 crc kubenswrapper[4948]: I0312 00:09:58.311757 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:09:58 crc kubenswrapper[4948]: I0312 00:09:58.311737 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:09:58 crc kubenswrapper[4948]: E0312 00:09:58.311859 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 00:09:58 crc kubenswrapper[4948]: E0312 00:09:58.312033 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 00:09:58 crc kubenswrapper[4948]: E0312 00:09:58.312195 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mqmtd" podUID="be985ff1-de8a-431f-a230-67894ebe52d5" Mar 12 00:09:58 crc kubenswrapper[4948]: E0312 00:09:58.312295 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 00:10:00 crc kubenswrapper[4948]: I0312 00:10:00.311771 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:10:00 crc kubenswrapper[4948]: I0312 00:10:00.311847 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:10:00 crc kubenswrapper[4948]: E0312 00:10:00.312015 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 00:10:00 crc kubenswrapper[4948]: I0312 00:10:00.312082 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:10:00 crc kubenswrapper[4948]: I0312 00:10:00.312149 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:10:00 crc kubenswrapper[4948]: E0312 00:10:00.312452 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mqmtd" podUID="be985ff1-de8a-431f-a230-67894ebe52d5" Mar 12 00:10:00 crc kubenswrapper[4948]: E0312 00:10:00.313183 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 00:10:00 crc kubenswrapper[4948]: E0312 00:10:00.313354 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 00:10:00 crc kubenswrapper[4948]: I0312 00:10:00.313834 4948 scope.go:117] "RemoveContainer" containerID="1d6a0a3fceb31bd31a6c978f0b3ac7ff768469ab3f311918e9dc72b424b00c27" Mar 12 00:10:01 crc kubenswrapper[4948]: I0312 00:10:01.091810 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f7mp5_15117613-9fad-48c7-98c4-a2d84502ded9/ovnkube-controller/3.log" Mar 12 00:10:01 crc kubenswrapper[4948]: I0312 00:10:01.096189 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" event={"ID":"15117613-9fad-48c7-98c4-a2d84502ded9","Type":"ContainerStarted","Data":"d3de0f7f4aa39ee28267d3a0041033e89da7e51713f417910e09007110b7b9c2"} Mar 12 00:10:01 crc kubenswrapper[4948]: I0312 00:10:01.097730 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:10:01 crc kubenswrapper[4948]: I0312 00:10:01.152411 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" podStartSLOduration=162.152387467 podStartE2EDuration="2m42.152387467s" podCreationTimestamp="2026-03-12 00:07:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:10:01.151025385 +0000 UTC m=+200.606629163" watchObservedRunningTime="2026-03-12 00:10:01.152387467 +0000 UTC m=+200.607991245" Mar 12 00:10:01 crc kubenswrapper[4948]: I0312 00:10:01.384379 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-mqmtd"] Mar 12 00:10:01 crc kubenswrapper[4948]: I0312 00:10:01.384541 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:10:01 crc kubenswrapper[4948]: E0312 00:10:01.384668 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mqmtd" podUID="be985ff1-de8a-431f-a230-67894ebe52d5" Mar 12 00:10:01 crc kubenswrapper[4948]: E0312 00:10:01.457123 4948 kubelet.go:2916] "Container runtime network not ready" networkReady="NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" Mar 12 00:10:02 crc kubenswrapper[4948]: I0312 00:10:02.313232 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:10:02 crc kubenswrapper[4948]: I0312 00:10:02.313270 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:10:02 crc kubenswrapper[4948]: I0312 00:10:02.313283 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:10:02 crc kubenswrapper[4948]: I0312 00:10:02.313251 4948 scope.go:117] "RemoveContainer" containerID="b28502ee0add02df45488b29f20d8be4102b0f3faaf8b5e26b5943a410163bf7" Mar 12 00:10:02 crc kubenswrapper[4948]: E0312 00:10:02.313499 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 00:10:02 crc kubenswrapper[4948]: E0312 00:10:02.313758 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 00:10:02 crc kubenswrapper[4948]: E0312 00:10:02.313904 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 00:10:03 crc kubenswrapper[4948]: I0312 00:10:03.105413 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-bwpqp_3b7580c2-5c23-4c67-807a-ea97a3df9398/kube-multus/1.log" Mar 12 00:10:03 crc kubenswrapper[4948]: I0312 00:10:03.105498 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-bwpqp" event={"ID":"3b7580c2-5c23-4c67-807a-ea97a3df9398","Type":"ContainerStarted","Data":"b283d9461d9f6b7694a42e106d5d1c94fa66752ee64832f229a2e6088e3fca81"} Mar 12 00:10:03 crc kubenswrapper[4948]: I0312 00:10:03.312174 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:10:03 crc kubenswrapper[4948]: E0312 00:10:03.312419 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mqmtd" podUID="be985ff1-de8a-431f-a230-67894ebe52d5" Mar 12 00:10:04 crc kubenswrapper[4948]: I0312 00:10:04.311506 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:10:04 crc kubenswrapper[4948]: I0312 00:10:04.311602 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:10:04 crc kubenswrapper[4948]: I0312 00:10:04.311653 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:10:04 crc kubenswrapper[4948]: E0312 00:10:04.311841 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 00:10:04 crc kubenswrapper[4948]: E0312 00:10:04.312249 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 00:10:04 crc kubenswrapper[4948]: E0312 00:10:04.312671 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 00:10:05 crc kubenswrapper[4948]: I0312 00:10:05.311796 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:10:05 crc kubenswrapper[4948]: E0312 00:10:05.312055 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-multus/network-metrics-daemon-mqmtd" podUID="be985ff1-de8a-431f-a230-67894ebe52d5" Mar 12 00:10:06 crc kubenswrapper[4948]: I0312 00:10:06.312034 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:10:06 crc kubenswrapper[4948]: E0312 00:10:06.313415 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" podUID="5fe485a1-e14f-4c09-b5b9-f252bc42b7e8" Mar 12 00:10:06 crc kubenswrapper[4948]: I0312 00:10:06.312046 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:10:06 crc kubenswrapper[4948]: I0312 00:10:06.312051 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:10:06 crc kubenswrapper[4948]: E0312 00:10:06.314193 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" podUID="9d751cbb-f2e2-430d-9754-c882a5e924a5" Mar 12 00:10:06 crc kubenswrapper[4948]: E0312 00:10:06.314035 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="network is not ready: container runtime network not ready: NetworkReady=false reason:NetworkPluginNotReady message:Network plugin returns error: no CNI configuration file in /etc/kubernetes/cni/net.d/. Has your network provider started?" pod="openshift-network-diagnostics/network-check-target-xd92c" podUID="3b6479f0-333b-4a96-9adf-2099afdc2447" Mar 12 00:10:07 crc kubenswrapper[4948]: I0312 00:10:07.312283 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:10:07 crc kubenswrapper[4948]: I0312 00:10:07.315049 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Mar 12 00:10:07 crc kubenswrapper[4948]: I0312 00:10:07.317214 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Mar 12 00:10:08 crc kubenswrapper[4948]: I0312 00:10:08.312077 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:10:08 crc kubenswrapper[4948]: I0312 00:10:08.312081 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:10:08 crc kubenswrapper[4948]: I0312 00:10:08.312098 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:10:08 crc kubenswrapper[4948]: I0312 00:10:08.315413 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Mar 12 00:10:08 crc kubenswrapper[4948]: I0312 00:10:08.315495 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Mar 12 00:10:08 crc kubenswrapper[4948]: I0312 00:10:08.316387 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Mar 12 00:10:08 crc kubenswrapper[4948]: I0312 00:10:08.316406 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Mar 12 00:10:14 crc kubenswrapper[4948]: I0312 00:10:14.784652 4948 patch_prober.go:28] interesting pod/machine-config-daemon-m4xwc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 00:10:14 crc kubenswrapper[4948]: I0312 00:10:14.786038 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" podUID="7e483c2b-08f0-4e92-8e4a-b7281f30af3e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 00:10:14 crc kubenswrapper[4948]: I0312 00:10:14.791905 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:10:14 crc kubenswrapper[4948]: I0312 00:10:14.973170 4948 kubelet_node_status.go:724] "Recording event message for node" node="crc" event="NodeReady" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.026118 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-g7g94"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.026995 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-g7g94" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.029237 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-9tpt4"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.030892 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-9tpt4" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.032120 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-fx7lk"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.032974 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fx7lk" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.040551 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vjnjc"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.041192 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.041442 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vjnjc" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.041993 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.042812 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.043184 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-rp8sv"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.043518 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.043762 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rp8sv" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.043881 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.044552 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.044634 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.044835 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.044994 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.053453 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-5cwjt"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.053550 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.053567 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.053638 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.053814 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.054408 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-5cwjt" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.054942 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-machine-approver/machine-approver-56656f9798-n4545"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.055666 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-n4545" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.056298 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-qsf8x"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.057200 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-qsf8x" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.059775 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-6jfc4"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.060400 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-6jfc4" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.067179 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.069472 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.069568 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.069607 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-sg87j"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.069792 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.069976 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.070119 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.069999 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.070038 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.070412 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-sg87j" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.072767 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.073022 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.073277 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.077953 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.078168 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.078231 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.078346 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.078497 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.078691 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.078801 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.078902 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.079376 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.079508 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.079571 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.079518 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.084932 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.085814 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.085977 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.088104 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.088456 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.088733 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.088955 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.089208 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.089429 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.089615 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.089881 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.090050 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.090232 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.090623 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.090846 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.107145 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.107329 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.108085 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-xs8b7"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.108919 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-xs8b7" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.110574 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4fn9j"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.111165 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4fn9j" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.111827 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.112008 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.112087 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.112177 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.112324 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.112442 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.112572 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.112712 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.113286 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.113419 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.113806 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.113846 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.114077 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.116288 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/downloads-7954f5f757-xdh6w"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.116659 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qc4wh"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.116939 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qc4wh" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.117173 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-xdh6w" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.121593 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.123280 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.123577 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.123764 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.121937 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.124442 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.127107 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.128540 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.133713 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.133959 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.134457 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.134619 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.134742 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.134988 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-pzgdk"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.135142 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.135486 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-pzgdk" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.135595 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.136115 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.136261 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.136403 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.136437 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console-operator/console-operator-58897d9998-z7w95"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.136477 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.136534 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.136405 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.136892 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-z7w95" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.137086 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.137175 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.137677 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-pruner-29554560-slpk9"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.138048 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29554560-slpk9" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.138433 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.139132 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rfxxc"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.139789 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rfxxc" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.147727 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6q7g6\" (UniqueName: \"kubernetes.io/projected/60024b0b-f817-477f-83f1-e53cbc1e05b0-kube-api-access-6q7g6\") pod \"route-controller-manager-6576b87f9c-rp8sv\" (UID: \"60024b0b-f817-477f-83f1-e53cbc1e05b0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rp8sv" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.147789 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f90934d6-9305-4513-8d84-cd6a6fcc4a59-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-qsf8x\" (UID: \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\") " pod="openshift-authentication/oauth-openshift-558db77b4-qsf8x" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.147826 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b36cdace-7dbb-4e8a-b0f0-6e93615612a8-config\") pod \"openshift-apiserver-operator-796bbdcf4f-vjnjc\" (UID: \"b36cdace-7dbb-4e8a-b0f0-6e93615612a8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vjnjc" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.147861 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f90934d6-9305-4513-8d84-cd6a6fcc4a59-audit-dir\") pod \"oauth-openshift-558db77b4-qsf8x\" (UID: \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\") " pod="openshift-authentication/oauth-openshift-558db77b4-qsf8x" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.147897 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dgm4m\" (UniqueName: \"kubernetes.io/projected/f90934d6-9305-4513-8d84-cd6a6fcc4a59-kube-api-access-dgm4m\") pod \"oauth-openshift-558db77b4-qsf8x\" (UID: \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\") " pod="openshift-authentication/oauth-openshift-558db77b4-qsf8x" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.147942 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f90934d6-9305-4513-8d84-cd6a6fcc4a59-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-qsf8x\" (UID: \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\") " pod="openshift-authentication/oauth-openshift-558db77b4-qsf8x" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.147975 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5c9bc3de-a78c-44d5-90ca-f57a625e125a-serving-cert\") pod \"authentication-operator-69f744f599-5cwjt\" (UID: \"5c9bc3de-a78c-44d5-90ca-f57a625e125a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5cwjt" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.148008 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-v9fgz\" (UniqueName: \"kubernetes.io/projected/f980811d-53fd-4c64-96c5-19283afcbb1f-kube-api-access-v9fgz\") pod \"apiserver-7bbb656c7d-fx7lk\" (UID: \"f980811d-53fd-4c64-96c5-19283afcbb1f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fx7lk" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.148049 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6da8984-cc2d-4b58-b6da-d7fc2d1e1725-config\") pod \"machine-approver-56656f9798-n4545\" (UID: \"a6da8984-cc2d-4b58-b6da-d7fc2d1e1725\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-n4545" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.148080 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f90934d6-9305-4513-8d84-cd6a6fcc4a59-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-qsf8x\" (UID: \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\") " pod="openshift-authentication/oauth-openshift-558db77b4-qsf8x" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.148107 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2040ed8d-ae32-4909-9161-bc7ffbe49edd-audit-dir\") pod \"apiserver-76f77b778f-6jfc4\" (UID: \"2040ed8d-ae32-4909-9161-bc7ffbe49edd\") " pod="openshift-apiserver/apiserver-76f77b778f-6jfc4" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.148146 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/a6da8984-cc2d-4b58-b6da-d7fc2d1e1725-machine-approver-tls\") pod \"machine-approver-56656f9798-n4545\" (UID: \"a6da8984-cc2d-4b58-b6da-d7fc2d1e1725\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-n4545" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.148183 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f90934d6-9305-4513-8d84-cd6a6fcc4a59-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-qsf8x\" (UID: \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\") " pod="openshift-authentication/oauth-openshift-558db77b4-qsf8x" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.148219 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f980811d-53fd-4c64-96c5-19283afcbb1f-etcd-client\") pod \"apiserver-7bbb656c7d-fx7lk\" (UID: \"f980811d-53fd-4c64-96c5-19283afcbb1f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fx7lk" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.148248 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f980811d-53fd-4c64-96c5-19283afcbb1f-serving-cert\") pod \"apiserver-7bbb656c7d-fx7lk\" (UID: \"f980811d-53fd-4c64-96c5-19283afcbb1f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fx7lk" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.148285 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/2040ed8d-ae32-4909-9161-bc7ffbe49edd-etcd-serving-ca\") pod \"apiserver-76f77b778f-6jfc4\" (UID: \"2040ed8d-ae32-4909-9161-bc7ffbe49edd\") " pod="openshift-apiserver/apiserver-76f77b778f-6jfc4" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.148382 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5c9bc3de-a78c-44d5-90ca-f57a625e125a-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-5cwjt\" (UID: \"5c9bc3de-a78c-44d5-90ca-f57a625e125a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5cwjt" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.148415 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f90934d6-9305-4513-8d84-cd6a6fcc4a59-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-qsf8x\" (UID: \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\") " pod="openshift-authentication/oauth-openshift-558db77b4-qsf8x" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.148447 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2040ed8d-ae32-4909-9161-bc7ffbe49edd-config\") pod \"apiserver-76f77b778f-6jfc4\" (UID: \"2040ed8d-ae32-4909-9161-bc7ffbe49edd\") " pod="openshift-apiserver/apiserver-76f77b778f-6jfc4" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.148482 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f980811d-53fd-4c64-96c5-19283afcbb1f-encryption-config\") pod \"apiserver-7bbb656c7d-fx7lk\" (UID: \"f980811d-53fd-4c64-96c5-19283afcbb1f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fx7lk" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.148534 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f980811d-53fd-4c64-96c5-19283afcbb1f-audit-dir\") pod \"apiserver-7bbb656c7d-fx7lk\" (UID: \"f980811d-53fd-4c64-96c5-19283afcbb1f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fx7lk" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.148563 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/60024b0b-f817-477f-83f1-e53cbc1e05b0-serving-cert\") pod \"route-controller-manager-6576b87f9c-rp8sv\" (UID: \"60024b0b-f817-477f-83f1-e53cbc1e05b0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rp8sv" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.148592 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2040ed8d-ae32-4909-9161-bc7ffbe49edd-serving-cert\") pod \"apiserver-76f77b778f-6jfc4\" (UID: \"2040ed8d-ae32-4909-9161-bc7ffbe49edd\") " pod="openshift-apiserver/apiserver-76f77b778f-6jfc4" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.148626 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-2j6ml\" (UniqueName: \"kubernetes.io/projected/9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5-kube-api-access-2j6ml\") pod \"controller-manager-879f6c89f-9tpt4\" (UID: \"9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9tpt4" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.148673 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-9tpt4\" (UID: \"9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9tpt4" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.148703 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/a544acdb-b758-4ea8-a43d-8270fc454901-images\") pod \"machine-api-operator-5694c8668f-g7g94\" (UID: \"a544acdb-b758-4ea8-a43d-8270fc454901\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-g7g94" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.148753 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f90934d6-9305-4513-8d84-cd6a6fcc4a59-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-qsf8x\" (UID: \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\") " pod="openshift-authentication/oauth-openshift-558db77b4-qsf8x" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.148785 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5c9bc3de-a78c-44d5-90ca-f57a625e125a-service-ca-bundle\") pod \"authentication-operator-69f744f599-5cwjt\" (UID: \"5c9bc3de-a78c-44d5-90ca-f57a625e125a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5cwjt" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.148809 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f90934d6-9305-4513-8d84-cd6a6fcc4a59-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-qsf8x\" (UID: \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\") " pod="openshift-authentication/oauth-openshift-558db77b4-qsf8x" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.148840 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2040ed8d-ae32-4909-9161-bc7ffbe49edd-etcd-client\") pod \"apiserver-76f77b778f-6jfc4\" (UID: \"2040ed8d-ae32-4909-9161-bc7ffbe49edd\") " pod="openshift-apiserver/apiserver-76f77b778f-6jfc4" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.148873 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f980811d-53fd-4c64-96c5-19283afcbb1f-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-fx7lk\" (UID: \"f980811d-53fd-4c64-96c5-19283afcbb1f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fx7lk" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.148909 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vrqxj\" (UniqueName: \"kubernetes.io/projected/b36cdace-7dbb-4e8a-b0f0-6e93615612a8-kube-api-access-vrqxj\") pod \"openshift-apiserver-operator-796bbdcf4f-vjnjc\" (UID: \"b36cdace-7dbb-4e8a-b0f0-6e93615612a8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vjnjc" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.148942 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a12c201c-f0ce-4519-8873-e7ec122e7fec-serving-cert\") pod \"openshift-config-operator-7777fb866f-sg87j\" (UID: \"a12c201c-f0ce-4519-8873-e7ec122e7fec\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-sg87j" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.148992 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a6da8984-cc2d-4b58-b6da-d7fc2d1e1725-auth-proxy-config\") pod \"machine-approver-56656f9798-n4545\" (UID: \"a6da8984-cc2d-4b58-b6da-d7fc2d1e1725\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-n4545" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.149026 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f90934d6-9305-4513-8d84-cd6a6fcc4a59-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-qsf8x\" (UID: \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\") " pod="openshift-authentication/oauth-openshift-558db77b4-qsf8x" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.149058 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/2040ed8d-ae32-4909-9161-bc7ffbe49edd-image-import-ca\") pod \"apiserver-76f77b778f-6jfc4\" (UID: \"2040ed8d-ae32-4909-9161-bc7ffbe49edd\") " pod="openshift-apiserver/apiserver-76f77b778f-6jfc4" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.149089 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gl6qb\" (UniqueName: \"kubernetes.io/projected/5c9bc3de-a78c-44d5-90ca-f57a625e125a-kube-api-access-gl6qb\") pod \"authentication-operator-69f744f599-5cwjt\" (UID: \"5c9bc3de-a78c-44d5-90ca-f57a625e125a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5cwjt" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.149126 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f90934d6-9305-4513-8d84-cd6a6fcc4a59-audit-policies\") pod \"oauth-openshift-558db77b4-qsf8x\" (UID: \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\") " pod="openshift-authentication/oauth-openshift-558db77b4-qsf8x" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.149155 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l8vtg\" (UniqueName: \"kubernetes.io/projected/2040ed8d-ae32-4909-9161-bc7ffbe49edd-kube-api-access-l8vtg\") pod \"apiserver-76f77b778f-6jfc4\" (UID: \"2040ed8d-ae32-4909-9161-bc7ffbe49edd\") " pod="openshift-apiserver/apiserver-76f77b778f-6jfc4" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.149189 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4xvzv\" (UniqueName: \"kubernetes.io/projected/a12c201c-f0ce-4519-8873-e7ec122e7fec-kube-api-access-4xvzv\") pod \"openshift-config-operator-7777fb866f-sg87j\" (UID: \"a12c201c-f0ce-4519-8873-e7ec122e7fec\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-sg87j" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.149214 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5-config\") pod \"controller-manager-879f6c89f-9tpt4\" (UID: \"9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9tpt4" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.149242 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/f980811d-53fd-4c64-96c5-19283afcbb1f-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-fx7lk\" (UID: \"f980811d-53fd-4c64-96c5-19283afcbb1f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fx7lk" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.149275 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f90934d6-9305-4513-8d84-cd6a6fcc4a59-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-qsf8x\" (UID: \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\") " pod="openshift-authentication/oauth-openshift-558db77b4-qsf8x" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.149334 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/2040ed8d-ae32-4909-9161-bc7ffbe49edd-node-pullsecrets\") pod \"apiserver-76f77b778f-6jfc4\" (UID: \"2040ed8d-ae32-4909-9161-bc7ffbe49edd\") " pod="openshift-apiserver/apiserver-76f77b778f-6jfc4" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.149366 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dcmgg\" (UniqueName: \"kubernetes.io/projected/a544acdb-b758-4ea8-a43d-8270fc454901-kube-api-access-dcmgg\") pod \"machine-api-operator-5694c8668f-g7g94\" (UID: \"a544acdb-b758-4ea8-a43d-8270fc454901\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-g7g94" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.149395 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2040ed8d-ae32-4909-9161-bc7ffbe49edd-trusted-ca-bundle\") pod \"apiserver-76f77b778f-6jfc4\" (UID: \"2040ed8d-ae32-4909-9161-bc7ffbe49edd\") " pod="openshift-apiserver/apiserver-76f77b778f-6jfc4" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.149443 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f90934d6-9305-4513-8d84-cd6a6fcc4a59-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-qsf8x\" (UID: \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\") " pod="openshift-authentication/oauth-openshift-558db77b4-qsf8x" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.149474 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/2040ed8d-ae32-4909-9161-bc7ffbe49edd-audit\") pod \"apiserver-76f77b778f-6jfc4\" (UID: \"2040ed8d-ae32-4909-9161-bc7ffbe49edd\") " pod="openshift-apiserver/apiserver-76f77b778f-6jfc4" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.149502 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60024b0b-f817-477f-83f1-e53cbc1e05b0-config\") pod \"route-controller-manager-6576b87f9c-rp8sv\" (UID: \"60024b0b-f817-477f-83f1-e53cbc1e05b0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rp8sv" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.149529 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/2040ed8d-ae32-4909-9161-bc7ffbe49edd-encryption-config\") pod \"apiserver-76f77b778f-6jfc4\" (UID: \"2040ed8d-ae32-4909-9161-bc7ffbe49edd\") " pod="openshift-apiserver/apiserver-76f77b778f-6jfc4" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.149559 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5-client-ca\") pod \"controller-manager-879f6c89f-9tpt4\" (UID: \"9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9tpt4" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.149587 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmk5h\" (UniqueName: \"kubernetes.io/projected/a6da8984-cc2d-4b58-b6da-d7fc2d1e1725-kube-api-access-fmk5h\") pod \"machine-approver-56656f9798-n4545\" (UID: \"a6da8984-cc2d-4b58-b6da-d7fc2d1e1725\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-n4545" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.149620 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c9bc3de-a78c-44d5-90ca-f57a625e125a-config\") pod \"authentication-operator-69f744f599-5cwjt\" (UID: \"5c9bc3de-a78c-44d5-90ca-f57a625e125a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5cwjt" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.149648 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b36cdace-7dbb-4e8a-b0f0-6e93615612a8-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-vjnjc\" (UID: \"b36cdace-7dbb-4e8a-b0f0-6e93615612a8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vjnjc" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.149679 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f90934d6-9305-4513-8d84-cd6a6fcc4a59-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-qsf8x\" (UID: \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\") " pod="openshift-authentication/oauth-openshift-558db77b4-qsf8x" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.149702 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f980811d-53fd-4c64-96c5-19283afcbb1f-audit-policies\") pod \"apiserver-7bbb656c7d-fx7lk\" (UID: \"f980811d-53fd-4c64-96c5-19283afcbb1f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fx7lk" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.149728 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/a12c201c-f0ce-4519-8873-e7ec122e7fec-available-featuregates\") pod \"openshift-config-operator-7777fb866f-sg87j\" (UID: \"a12c201c-f0ce-4519-8873-e7ec122e7fec\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-sg87j" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.149756 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/a544acdb-b758-4ea8-a43d-8270fc454901-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-g7g94\" (UID: \"a544acdb-b758-4ea8-a43d-8270fc454901\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-g7g94" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.149791 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a544acdb-b758-4ea8-a43d-8270fc454901-config\") pod \"machine-api-operator-5694c8668f-g7g94\" (UID: \"a544acdb-b758-4ea8-a43d-8270fc454901\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-g7g94" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.149819 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5-serving-cert\") pod \"controller-manager-879f6c89f-9tpt4\" (UID: \"9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9tpt4" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.149864 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/60024b0b-f817-477f-83f1-e53cbc1e05b0-client-ca\") pod \"route-controller-manager-6576b87f9c-rp8sv\" (UID: \"60024b0b-f817-477f-83f1-e53cbc1e05b0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rp8sv" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.157029 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.157574 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.157773 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.158427 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.158627 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.159152 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.161295 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.161678 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-console/console-f9d7485db-qq8gg"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.165715 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.166020 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.166267 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.166746 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.187504 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-z6v9h"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.187951 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.187996 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-g7g94"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.188011 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kvh9w"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.188262 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-qq8gg" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.188320 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-srr9m"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.188676 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress/router-default-5444994796-7vhkr"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.188921 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6m4k7"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.189272 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6m4k7" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.189704 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-z6v9h" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.189749 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-7vhkr" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.189804 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kvh9w" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.189961 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-srr9m" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.189706 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bj72v"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.191530 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gc7vl"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.191829 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-d24n2"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.191990 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bj72v" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.192068 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gc7vl" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.192680 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-9k744"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.193106 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9k744" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.193137 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-d24n2" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.194132 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.194551 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"serviceca" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.196708 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-zkbhv"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.197317 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-zkbhv" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.197660 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-g4kcb"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.198164 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.200004 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-kjggj"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.200583 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-kjggj" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.200942 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h6z8v"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.201334 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h6z8v" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.203888 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jlrln"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.204618 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29554560-8mgv9"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.205043 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29554570-2b2hb"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.205230 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29554560-8mgv9" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.205506 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jlrln" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.205682 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29554570-2b2hb" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.206702 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5qjdn"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.207291 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5qjdn" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.207783 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-v59sv"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.208609 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-v59sv" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.208834 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dc6wb"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.209329 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dc6wb" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.211427 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"pruner-dockercfg-p7bcw" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.214145 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-fx7lk"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.215236 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vjnjc"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.216175 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-9tpt4"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.217138 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-lkhdg"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.218251 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-q79m4"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.218382 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lkhdg" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.219042 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-q79m4" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.219168 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-rp8sv"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.220088 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ingress-canary/ingress-canary-mdhf9"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.220655 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-mdhf9" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.223071 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-5cwjt"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.224172 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kvh9w"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.225353 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rfxxc"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.228136 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qc4wh"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.228997 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-sg87j"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.230240 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-6jfc4"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.231358 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-xdh6w"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.233540 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-pzgdk"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.234540 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.235281 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-qsf8x"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.236410 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-qq8gg"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.237495 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bj72v"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.238784 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6m4k7"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.240059 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-srr9m"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.241453 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h6z8v"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.250268 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.253955 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/2040ed8d-ae32-4909-9161-bc7ffbe49edd-image-import-ca\") pod \"apiserver-76f77b778f-6jfc4\" (UID: \"2040ed8d-ae32-4909-9161-bc7ffbe49edd\") " pod="openshift-apiserver/apiserver-76f77b778f-6jfc4" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.254000 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gl6qb\" (UniqueName: \"kubernetes.io/projected/5c9bc3de-a78c-44d5-90ca-f57a625e125a-kube-api-access-gl6qb\") pod \"authentication-operator-69f744f599-5cwjt\" (UID: \"5c9bc3de-a78c-44d5-90ca-f57a625e125a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5cwjt" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.254029 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f90934d6-9305-4513-8d84-cd6a6fcc4a59-audit-policies\") pod \"oauth-openshift-558db77b4-qsf8x\" (UID: \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\") " pod="openshift-authentication/oauth-openshift-558db77b4-qsf8x" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.254051 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l8vtg\" (UniqueName: \"kubernetes.io/projected/2040ed8d-ae32-4909-9161-bc7ffbe49edd-kube-api-access-l8vtg\") pod \"apiserver-76f77b778f-6jfc4\" (UID: \"2040ed8d-ae32-4909-9161-bc7ffbe49edd\") " pod="openshift-apiserver/apiserver-76f77b778f-6jfc4" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.254071 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4xvzv\" (UniqueName: \"kubernetes.io/projected/a12c201c-f0ce-4519-8873-e7ec122e7fec-kube-api-access-4xvzv\") pod \"openshift-config-operator-7777fb866f-sg87j\" (UID: \"a12c201c-f0ce-4519-8873-e7ec122e7fec\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-sg87j" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.254087 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5-config\") pod \"controller-manager-879f6c89f-9tpt4\" (UID: \"9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9tpt4" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.254106 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/f980811d-53fd-4c64-96c5-19283afcbb1f-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-fx7lk\" (UID: \"f980811d-53fd-4c64-96c5-19283afcbb1f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fx7lk" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.254129 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f90934d6-9305-4513-8d84-cd6a6fcc4a59-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-qsf8x\" (UID: \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\") " pod="openshift-authentication/oauth-openshift-558db77b4-qsf8x" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.254149 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/2040ed8d-ae32-4909-9161-bc7ffbe49edd-node-pullsecrets\") pod \"apiserver-76f77b778f-6jfc4\" (UID: \"2040ed8d-ae32-4909-9161-bc7ffbe49edd\") " pod="openshift-apiserver/apiserver-76f77b778f-6jfc4" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.254176 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dcmgg\" (UniqueName: \"kubernetes.io/projected/a544acdb-b758-4ea8-a43d-8270fc454901-kube-api-access-dcmgg\") pod \"machine-api-operator-5694c8668f-g7g94\" (UID: \"a544acdb-b758-4ea8-a43d-8270fc454901\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-g7g94" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.254422 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2040ed8d-ae32-4909-9161-bc7ffbe49edd-trusted-ca-bundle\") pod \"apiserver-76f77b778f-6jfc4\" (UID: \"2040ed8d-ae32-4909-9161-bc7ffbe49edd\") " pod="openshift-apiserver/apiserver-76f77b778f-6jfc4" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.254564 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f90934d6-9305-4513-8d84-cd6a6fcc4a59-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-qsf8x\" (UID: \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\") " pod="openshift-authentication/oauth-openshift-558db77b4-qsf8x" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.255588 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/2040ed8d-ae32-4909-9161-bc7ffbe49edd-audit\") pod \"apiserver-76f77b778f-6jfc4\" (UID: \"2040ed8d-ae32-4909-9161-bc7ffbe49edd\") " pod="openshift-apiserver/apiserver-76f77b778f-6jfc4" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.255868 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-pullsecrets\" (UniqueName: \"kubernetes.io/host-path/2040ed8d-ae32-4909-9161-bc7ffbe49edd-node-pullsecrets\") pod \"apiserver-76f77b778f-6jfc4\" (UID: \"2040ed8d-ae32-4909-9161-bc7ffbe49edd\") " pod="openshift-apiserver/apiserver-76f77b778f-6jfc4" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.256839 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/f980811d-53fd-4c64-96c5-19283afcbb1f-etcd-serving-ca\") pod \"apiserver-7bbb656c7d-fx7lk\" (UID: \"f980811d-53fd-4c64-96c5-19283afcbb1f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fx7lk" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.257088 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29554560-8mgv9"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.257089 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5-config\") pod \"controller-manager-879f6c89f-9tpt4\" (UID: \"9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9tpt4" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.258111 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit\" (UniqueName: \"kubernetes.io/configmap/2040ed8d-ae32-4909-9161-bc7ffbe49edd-audit\") pod \"apiserver-76f77b778f-6jfc4\" (UID: \"2040ed8d-ae32-4909-9161-bc7ffbe49edd\") " pod="openshift-apiserver/apiserver-76f77b778f-6jfc4" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.258759 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60024b0b-f817-477f-83f1-e53cbc1e05b0-config\") pod \"route-controller-manager-6576b87f9c-rp8sv\" (UID: \"60024b0b-f817-477f-83f1-e53cbc1e05b0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rp8sv" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.258577 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f90934d6-9305-4513-8d84-cd6a6fcc4a59-audit-policies\") pod \"oauth-openshift-558db77b4-qsf8x\" (UID: \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\") " pod="openshift-authentication/oauth-openshift-558db77b4-qsf8x" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.258687 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-import-ca\" (UniqueName: \"kubernetes.io/configmap/2040ed8d-ae32-4909-9161-bc7ffbe49edd-image-import-ca\") pod \"apiserver-76f77b778f-6jfc4\" (UID: \"2040ed8d-ae32-4909-9161-bc7ffbe49edd\") " pod="openshift-apiserver/apiserver-76f77b778f-6jfc4" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.258359 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/2040ed8d-ae32-4909-9161-bc7ffbe49edd-trusted-ca-bundle\") pod \"apiserver-76f77b778f-6jfc4\" (UID: \"2040ed8d-ae32-4909-9161-bc7ffbe49edd\") " pod="openshift-apiserver/apiserver-76f77b778f-6jfc4" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.258861 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/2040ed8d-ae32-4909-9161-bc7ffbe49edd-encryption-config\") pod \"apiserver-76f77b778f-6jfc4\" (UID: \"2040ed8d-ae32-4909-9161-bc7ffbe49edd\") " pod="openshift-apiserver/apiserver-76f77b778f-6jfc4" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.258901 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5-client-ca\") pod \"controller-manager-879f6c89f-9tpt4\" (UID: \"9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9tpt4" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.258993 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmk5h\" (UniqueName: \"kubernetes.io/projected/a6da8984-cc2d-4b58-b6da-d7fc2d1e1725-kube-api-access-fmk5h\") pod \"machine-approver-56656f9798-n4545\" (UID: \"a6da8984-cc2d-4b58-b6da-d7fc2d1e1725\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-n4545" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.259045 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c9bc3de-a78c-44d5-90ca-f57a625e125a-config\") pod \"authentication-operator-69f744f599-5cwjt\" (UID: \"5c9bc3de-a78c-44d5-90ca-f57a625e125a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5cwjt" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.259109 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b36cdace-7dbb-4e8a-b0f0-6e93615612a8-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-vjnjc\" (UID: \"b36cdace-7dbb-4e8a-b0f0-6e93615612a8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vjnjc" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.259151 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f90934d6-9305-4513-8d84-cd6a6fcc4a59-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-qsf8x\" (UID: \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\") " pod="openshift-authentication/oauth-openshift-558db77b4-qsf8x" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.260005 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f90934d6-9305-4513-8d84-cd6a6fcc4a59-v4-0-config-system-cliconfig\") pod \"oauth-openshift-558db77b4-qsf8x\" (UID: \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\") " pod="openshift-authentication/oauth-openshift-558db77b4-qsf8x" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.260103 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5c9bc3de-a78c-44d5-90ca-f57a625e125a-config\") pod \"authentication-operator-69f744f599-5cwjt\" (UID: \"5c9bc3de-a78c-44d5-90ca-f57a625e125a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5cwjt" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.260421 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-g4kcb"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.260748 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60024b0b-f817-477f-83f1-e53cbc1e05b0-config\") pod \"route-controller-manager-6576b87f9c-rp8sv\" (UID: \"60024b0b-f817-477f-83f1-e53cbc1e05b0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rp8sv" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.260848 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f980811d-53fd-4c64-96c5-19283afcbb1f-audit-policies\") pod \"apiserver-7bbb656c7d-fx7lk\" (UID: \"f980811d-53fd-4c64-96c5-19283afcbb1f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fx7lk" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.260916 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/a12c201c-f0ce-4519-8873-e7ec122e7fec-available-featuregates\") pod \"openshift-config-operator-7777fb866f-sg87j\" (UID: \"a12c201c-f0ce-4519-8873-e7ec122e7fec\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-sg87j" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.260952 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/a544acdb-b758-4ea8-a43d-8270fc454901-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-g7g94\" (UID: \"a544acdb-b758-4ea8-a43d-8270fc454901\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-g7g94" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.261002 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t9zdh\" (UniqueName: \"kubernetes.io/projected/74ab81d8-55e8-47da-adb5-1f616a100a31-kube-api-access-t9zdh\") pod \"downloads-7954f5f757-xdh6w\" (UID: \"74ab81d8-55e8-47da-adb5-1f616a100a31\") " pod="openshift-console/downloads-7954f5f757-xdh6w" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.261029 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a544acdb-b758-4ea8-a43d-8270fc454901-config\") pod \"machine-api-operator-5694c8668f-g7g94\" (UID: \"a544acdb-b758-4ea8-a43d-8270fc454901\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-g7g94" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.261069 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5-serving-cert\") pod \"controller-manager-879f6c89f-9tpt4\" (UID: \"9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9tpt4" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.261110 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/60024b0b-f817-477f-83f1-e53cbc1e05b0-client-ca\") pod \"route-controller-manager-6576b87f9c-rp8sv\" (UID: \"60024b0b-f817-477f-83f1-e53cbc1e05b0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rp8sv" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.261464 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f90934d6-9305-4513-8d84-cd6a6fcc4a59-v4-0-config-system-service-ca\") pod \"oauth-openshift-558db77b4-qsf8x\" (UID: \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\") " pod="openshift-authentication/oauth-openshift-558db77b4-qsf8x" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.261685 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"available-featuregates\" (UniqueName: \"kubernetes.io/empty-dir/a12c201c-f0ce-4519-8873-e7ec122e7fec-available-featuregates\") pod \"openshift-config-operator-7777fb866f-sg87j\" (UID: \"a12c201c-f0ce-4519-8873-e7ec122e7fec\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-sg87j" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.261151 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6q7g6\" (UniqueName: \"kubernetes.io/projected/60024b0b-f817-477f-83f1-e53cbc1e05b0-kube-api-access-6q7g6\") pod \"route-controller-manager-6576b87f9c-rp8sv\" (UID: \"60024b0b-f817-477f-83f1-e53cbc1e05b0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rp8sv" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.261759 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f90934d6-9305-4513-8d84-cd6a6fcc4a59-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-qsf8x\" (UID: \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\") " pod="openshift-authentication/oauth-openshift-558db77b4-qsf8x" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.261784 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b36cdace-7dbb-4e8a-b0f0-6e93615612a8-config\") pod \"openshift-apiserver-operator-796bbdcf4f-vjnjc\" (UID: \"b36cdace-7dbb-4e8a-b0f0-6e93615612a8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vjnjc" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.261809 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f90934d6-9305-4513-8d84-cd6a6fcc4a59-audit-dir\") pod \"oauth-openshift-558db77b4-qsf8x\" (UID: \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\") " pod="openshift-authentication/oauth-openshift-558db77b4-qsf8x" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.261829 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dgm4m\" (UniqueName: \"kubernetes.io/projected/f90934d6-9305-4513-8d84-cd6a6fcc4a59-kube-api-access-dgm4m\") pod \"oauth-openshift-558db77b4-qsf8x\" (UID: \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\") " pod="openshift-authentication/oauth-openshift-558db77b4-qsf8x" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.261856 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f90934d6-9305-4513-8d84-cd6a6fcc4a59-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-qsf8x\" (UID: \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\") " pod="openshift-authentication/oauth-openshift-558db77b4-qsf8x" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.261876 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5c9bc3de-a78c-44d5-90ca-f57a625e125a-serving-cert\") pod \"authentication-operator-69f744f599-5cwjt\" (UID: \"5c9bc3de-a78c-44d5-90ca-f57a625e125a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5cwjt" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.261897 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-v9fgz\" (UniqueName: \"kubernetes.io/projected/f980811d-53fd-4c64-96c5-19283afcbb1f-kube-api-access-v9fgz\") pod \"apiserver-7bbb656c7d-fx7lk\" (UID: \"f980811d-53fd-4c64-96c5-19283afcbb1f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fx7lk" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.261919 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6da8984-cc2d-4b58-b6da-d7fc2d1e1725-config\") pod \"machine-approver-56656f9798-n4545\" (UID: \"a6da8984-cc2d-4b58-b6da-d7fc2d1e1725\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-n4545" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.261935 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f90934d6-9305-4513-8d84-cd6a6fcc4a59-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-qsf8x\" (UID: \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\") " pod="openshift-authentication/oauth-openshift-558db77b4-qsf8x" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.261953 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2040ed8d-ae32-4909-9161-bc7ffbe49edd-audit-dir\") pod \"apiserver-76f77b778f-6jfc4\" (UID: \"2040ed8d-ae32-4909-9161-bc7ffbe49edd\") " pod="openshift-apiserver/apiserver-76f77b778f-6jfc4" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.261973 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/a6da8984-cc2d-4b58-b6da-d7fc2d1e1725-machine-approver-tls\") pod \"machine-approver-56656f9798-n4545\" (UID: \"a6da8984-cc2d-4b58-b6da-d7fc2d1e1725\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-n4545" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.261995 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f90934d6-9305-4513-8d84-cd6a6fcc4a59-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-qsf8x\" (UID: \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\") " pod="openshift-authentication/oauth-openshift-558db77b4-qsf8x" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.262017 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f980811d-53fd-4c64-96c5-19283afcbb1f-etcd-client\") pod \"apiserver-7bbb656c7d-fx7lk\" (UID: \"f980811d-53fd-4c64-96c5-19283afcbb1f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fx7lk" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.262034 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f980811d-53fd-4c64-96c5-19283afcbb1f-serving-cert\") pod \"apiserver-7bbb656c7d-fx7lk\" (UID: \"f980811d-53fd-4c64-96c5-19283afcbb1f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fx7lk" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.262053 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/2040ed8d-ae32-4909-9161-bc7ffbe49edd-etcd-serving-ca\") pod \"apiserver-76f77b778f-6jfc4\" (UID: \"2040ed8d-ae32-4909-9161-bc7ffbe49edd\") " pod="openshift-apiserver/apiserver-76f77b778f-6jfc4" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.262086 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5c9bc3de-a78c-44d5-90ca-f57a625e125a-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-5cwjt\" (UID: \"5c9bc3de-a78c-44d5-90ca-f57a625e125a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5cwjt" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.262108 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f90934d6-9305-4513-8d84-cd6a6fcc4a59-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-qsf8x\" (UID: \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\") " pod="openshift-authentication/oauth-openshift-558db77b4-qsf8x" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.262127 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2040ed8d-ae32-4909-9161-bc7ffbe49edd-config\") pod \"apiserver-76f77b778f-6jfc4\" (UID: \"2040ed8d-ae32-4909-9161-bc7ffbe49edd\") " pod="openshift-apiserver/apiserver-76f77b778f-6jfc4" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.262145 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f980811d-53fd-4c64-96c5-19283afcbb1f-encryption-config\") pod \"apiserver-7bbb656c7d-fx7lk\" (UID: \"f980811d-53fd-4c64-96c5-19283afcbb1f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fx7lk" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.262162 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f980811d-53fd-4c64-96c5-19283afcbb1f-audit-dir\") pod \"apiserver-7bbb656c7d-fx7lk\" (UID: \"f980811d-53fd-4c64-96c5-19283afcbb1f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fx7lk" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.262181 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/60024b0b-f817-477f-83f1-e53cbc1e05b0-serving-cert\") pod \"route-controller-manager-6576b87f9c-rp8sv\" (UID: \"60024b0b-f817-477f-83f1-e53cbc1e05b0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rp8sv" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.262198 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2040ed8d-ae32-4909-9161-bc7ffbe49edd-serving-cert\") pod \"apiserver-76f77b778f-6jfc4\" (UID: \"2040ed8d-ae32-4909-9161-bc7ffbe49edd\") " pod="openshift-apiserver/apiserver-76f77b778f-6jfc4" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.262183 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f980811d-53fd-4c64-96c5-19283afcbb1f-audit-policies\") pod \"apiserver-7bbb656c7d-fx7lk\" (UID: \"f980811d-53fd-4c64-96c5-19283afcbb1f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fx7lk" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.262218 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-2j6ml\" (UniqueName: \"kubernetes.io/projected/9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5-kube-api-access-2j6ml\") pod \"controller-manager-879f6c89f-9tpt4\" (UID: \"9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9tpt4" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.262477 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-9tpt4\" (UID: \"9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9tpt4" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.262511 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/a544acdb-b758-4ea8-a43d-8270fc454901-images\") pod \"machine-api-operator-5694c8668f-g7g94\" (UID: \"a544acdb-b758-4ea8-a43d-8270fc454901\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-g7g94" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.262556 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f90934d6-9305-4513-8d84-cd6a6fcc4a59-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-qsf8x\" (UID: \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\") " pod="openshift-authentication/oauth-openshift-558db77b4-qsf8x" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.262590 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5c9bc3de-a78c-44d5-90ca-f57a625e125a-service-ca-bundle\") pod \"authentication-operator-69f744f599-5cwjt\" (UID: \"5c9bc3de-a78c-44d5-90ca-f57a625e125a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5cwjt" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.262617 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f90934d6-9305-4513-8d84-cd6a6fcc4a59-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-qsf8x\" (UID: \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\") " pod="openshift-authentication/oauth-openshift-558db77b4-qsf8x" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.262647 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2040ed8d-ae32-4909-9161-bc7ffbe49edd-etcd-client\") pod \"apiserver-76f77b778f-6jfc4\" (UID: \"2040ed8d-ae32-4909-9161-bc7ffbe49edd\") " pod="openshift-apiserver/apiserver-76f77b778f-6jfc4" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.262680 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f980811d-53fd-4c64-96c5-19283afcbb1f-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-fx7lk\" (UID: \"f980811d-53fd-4c64-96c5-19283afcbb1f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fx7lk" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.262709 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vrqxj\" (UniqueName: \"kubernetes.io/projected/b36cdace-7dbb-4e8a-b0f0-6e93615612a8-kube-api-access-vrqxj\") pod \"openshift-apiserver-operator-796bbdcf4f-vjnjc\" (UID: \"b36cdace-7dbb-4e8a-b0f0-6e93615612a8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vjnjc" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.262740 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a12c201c-f0ce-4519-8873-e7ec122e7fec-serving-cert\") pod \"openshift-config-operator-7777fb866f-sg87j\" (UID: \"a12c201c-f0ce-4519-8873-e7ec122e7fec\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-sg87j" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.262789 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a6da8984-cc2d-4b58-b6da-d7fc2d1e1725-auth-proxy-config\") pod \"machine-approver-56656f9798-n4545\" (UID: \"a6da8984-cc2d-4b58-b6da-d7fc2d1e1725\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-n4545" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.262815 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f90934d6-9305-4513-8d84-cd6a6fcc4a59-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-qsf8x\" (UID: \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\") " pod="openshift-authentication/oauth-openshift-558db77b4-qsf8x" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.264321 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-d24n2"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.264528 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b36cdace-7dbb-4e8a-b0f0-6e93615612a8-config\") pod \"openshift-apiserver-operator-796bbdcf4f-vjnjc\" (UID: \"b36cdace-7dbb-4e8a-b0f0-6e93615612a8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vjnjc" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.264586 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f90934d6-9305-4513-8d84-cd6a6fcc4a59-audit-dir\") pod \"oauth-openshift-558db77b4-qsf8x\" (UID: \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\") " pod="openshift-authentication/oauth-openshift-558db77b4-qsf8x" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.264587 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5-client-ca\") pod \"controller-manager-879f6c89f-9tpt4\" (UID: \"9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9tpt4" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.264826 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f980811d-53fd-4c64-96c5-19283afcbb1f-trusted-ca-bundle\") pod \"apiserver-7bbb656c7d-fx7lk\" (UID: \"f980811d-53fd-4c64-96c5-19283afcbb1f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fx7lk" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.265778 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/a6da8984-cc2d-4b58-b6da-d7fc2d1e1725-auth-proxy-config\") pod \"machine-approver-56656f9798-n4545\" (UID: \"a6da8984-cc2d-4b58-b6da-d7fc2d1e1725\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-n4545" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.265820 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/60024b0b-f817-477f-83f1-e53cbc1e05b0-client-ca\") pod \"route-controller-manager-6576b87f9c-rp8sv\" (UID: \"60024b0b-f817-477f-83f1-e53cbc1e05b0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rp8sv" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.265865 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5-proxy-ca-bundles\") pod \"controller-manager-879f6c89f-9tpt4\" (UID: \"9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9tpt4" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.266291 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-serving-ca\" (UniqueName: \"kubernetes.io/configmap/2040ed8d-ae32-4909-9161-bc7ffbe49edd-etcd-serving-ca\") pod \"apiserver-76f77b778f-6jfc4\" (UID: \"2040ed8d-ae32-4909-9161-bc7ffbe49edd\") " pod="openshift-apiserver/apiserver-76f77b778f-6jfc4" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.266644 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a544acdb-b758-4ea8-a43d-8270fc454901-config\") pod \"machine-api-operator-5694c8668f-g7g94\" (UID: \"a544acdb-b758-4ea8-a43d-8270fc454901\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-g7g94" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.267001 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5c9bc3de-a78c-44d5-90ca-f57a625e125a-trusted-ca-bundle\") pod \"authentication-operator-69f744f599-5cwjt\" (UID: \"5c9bc3de-a78c-44d5-90ca-f57a625e125a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5cwjt" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.267934 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5-serving-cert\") pod \"controller-manager-879f6c89f-9tpt4\" (UID: \"9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9tpt4" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.268900 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a12c201c-f0ce-4519-8873-e7ec122e7fec-serving-cert\") pod \"openshift-config-operator-7777fb866f-sg87j\" (UID: \"a12c201c-f0ce-4519-8873-e7ec122e7fec\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-sg87j" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.268977 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f980811d-53fd-4c64-96c5-19283afcbb1f-audit-dir\") pod \"apiserver-7bbb656c7d-fx7lk\" (UID: \"f980811d-53fd-4c64-96c5-19283afcbb1f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fx7lk" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.269527 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/2040ed8d-ae32-4909-9161-bc7ffbe49edd-config\") pod \"apiserver-76f77b778f-6jfc4\" (UID: \"2040ed8d-ae32-4909-9161-bc7ffbe49edd\") " pod="openshift-apiserver/apiserver-76f77b778f-6jfc4" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.269741 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-api-operator-tls\" (UniqueName: \"kubernetes.io/secret/a544acdb-b758-4ea8-a43d-8270fc454901-machine-api-operator-tls\") pod \"machine-api-operator-5694c8668f-g7g94\" (UID: \"a544acdb-b758-4ea8-a43d-8270fc454901\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-g7g94" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.270156 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b36cdace-7dbb-4e8a-b0f0-6e93615612a8-serving-cert\") pod \"openshift-apiserver-operator-796bbdcf4f-vjnjc\" (UID: \"b36cdace-7dbb-4e8a-b0f0-6e93615612a8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vjnjc" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.270543 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f90934d6-9305-4513-8d84-cd6a6fcc4a59-v4-0-config-system-serving-cert\") pod \"oauth-openshift-558db77b4-qsf8x\" (UID: \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\") " pod="openshift-authentication/oauth-openshift-558db77b4-qsf8x" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.270562 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/5c9bc3de-a78c-44d5-90ca-f57a625e125a-service-ca-bundle\") pod \"authentication-operator-69f744f599-5cwjt\" (UID: \"5c9bc3de-a78c-44d5-90ca-f57a625e125a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5cwjt" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.270796 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/2040ed8d-ae32-4909-9161-bc7ffbe49edd-encryption-config\") pod \"apiserver-76f77b778f-6jfc4\" (UID: \"2040ed8d-ae32-4909-9161-bc7ffbe49edd\") " pod="openshift-apiserver/apiserver-76f77b778f-6jfc4" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.270814 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/a544acdb-b758-4ea8-a43d-8270fc454901-images\") pod \"machine-api-operator-5694c8668f-g7g94\" (UID: \"a544acdb-b758-4ea8-a43d-8270fc454901\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-g7g94" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.270857 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-z6v9h"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.271492 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4fn9j"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.271894 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a6da8984-cc2d-4b58-b6da-d7fc2d1e1725-config\") pod \"machine-approver-56656f9798-n4545\" (UID: \"a6da8984-cc2d-4b58-b6da-d7fc2d1e1725\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-n4545" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.272029 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f90934d6-9305-4513-8d84-cd6a6fcc4a59-v4-0-config-system-session\") pod \"oauth-openshift-558db77b4-qsf8x\" (UID: \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\") " pod="openshift-authentication/oauth-openshift-558db77b4-qsf8x" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.272059 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/2040ed8d-ae32-4909-9161-bc7ffbe49edd-audit-dir\") pod \"apiserver-76f77b778f-6jfc4\" (UID: \"2040ed8d-ae32-4909-9161-bc7ffbe49edd\") " pod="openshift-apiserver/apiserver-76f77b778f-6jfc4" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.272342 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/60024b0b-f817-477f-83f1-e53cbc1e05b0-serving-cert\") pod \"route-controller-manager-6576b87f9c-rp8sv\" (UID: \"60024b0b-f817-477f-83f1-e53cbc1e05b0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rp8sv" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.272458 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.272997 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f90934d6-9305-4513-8d84-cd6a6fcc4a59-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-558db77b4-qsf8x\" (UID: \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\") " pod="openshift-authentication/oauth-openshift-558db77b4-qsf8x" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.274019 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f90934d6-9305-4513-8d84-cd6a6fcc4a59-v4-0-config-user-template-login\") pod \"oauth-openshift-558db77b4-qsf8x\" (UID: \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\") " pod="openshift-authentication/oauth-openshift-558db77b4-qsf8x" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.274198 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"encryption-config\" (UniqueName: \"kubernetes.io/secret/f980811d-53fd-4c64-96c5-19283afcbb1f-encryption-config\") pod \"apiserver-7bbb656c7d-fx7lk\" (UID: \"f980811d-53fd-4c64-96c5-19283afcbb1f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fx7lk" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.274411 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f90934d6-9305-4513-8d84-cd6a6fcc4a59-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-558db77b4-qsf8x\" (UID: \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\") " pod="openshift-authentication/oauth-openshift-558db77b4-qsf8x" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.274723 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"machine-approver-tls\" (UniqueName: \"kubernetes.io/secret/a6da8984-cc2d-4b58-b6da-d7fc2d1e1725-machine-approver-tls\") pod \"machine-approver-56656f9798-n4545\" (UID: \"a6da8984-cc2d-4b58-b6da-d7fc2d1e1725\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-n4545" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.274762 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f90934d6-9305-4513-8d84-cd6a6fcc4a59-v4-0-config-user-template-error\") pod \"oauth-openshift-558db77b4-qsf8x\" (UID: \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\") " pod="openshift-authentication/oauth-openshift-558db77b4-qsf8x" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.275238 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/2040ed8d-ae32-4909-9161-bc7ffbe49edd-etcd-client\") pod \"apiserver-76f77b778f-6jfc4\" (UID: \"2040ed8d-ae32-4909-9161-bc7ffbe49edd\") " pod="openshift-apiserver/apiserver-76f77b778f-6jfc4" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.275352 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/2040ed8d-ae32-4909-9161-bc7ffbe49edd-serving-cert\") pod \"apiserver-76f77b778f-6jfc4\" (UID: \"2040ed8d-ae32-4909-9161-bc7ffbe49edd\") " pod="openshift-apiserver/apiserver-76f77b778f-6jfc4" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.280712 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f90934d6-9305-4513-8d84-cd6a6fcc4a59-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-558db77b4-qsf8x\" (UID: \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\") " pod="openshift-authentication/oauth-openshift-558db77b4-qsf8x" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.287271 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f90934d6-9305-4513-8d84-cd6a6fcc4a59-v4-0-config-system-router-certs\") pod \"oauth-openshift-558db77b4-qsf8x\" (UID: \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\") " pod="openshift-authentication/oauth-openshift-558db77b4-qsf8x" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.291844 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/f980811d-53fd-4c64-96c5-19283afcbb1f-etcd-client\") pod \"apiserver-7bbb656c7d-fx7lk\" (UID: \"f980811d-53fd-4c64-96c5-19283afcbb1f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fx7lk" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.293555 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.293567 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f90934d6-9305-4513-8d84-cd6a6fcc4a59-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-558db77b4-qsf8x\" (UID: \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\") " pod="openshift-authentication/oauth-openshift-558db77b4-qsf8x" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.293939 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/f980811d-53fd-4c64-96c5-19283afcbb1f-serving-cert\") pod \"apiserver-7bbb656c7d-fx7lk\" (UID: \"f980811d-53fd-4c64-96c5-19283afcbb1f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fx7lk" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.295913 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5c9bc3de-a78c-44d5-90ca-f57a625e125a-serving-cert\") pod \"authentication-operator-69f744f599-5cwjt\" (UID: \"5c9bc3de-a78c-44d5-90ca-f57a625e125a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5cwjt" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.300912 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dc6wb"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.307866 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-q79m4"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.308703 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.309074 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-z7w95"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.310604 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-pruner-29554560-slpk9"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.318789 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-xs8b7"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.318821 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5qjdn"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.318831 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29554570-2b2hb"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.318842 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-9k744"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.318851 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jlrln"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.318861 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-dns/dns-default-prqnt"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.319457 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gc7vl"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.319535 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-prqnt" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.319859 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-kjggj"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.320871 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-5q9rs"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.322187 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-zkbhv"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.322267 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-5q9rs" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.323010 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-v59sv"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.323984 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-lkhdg"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.325006 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-prqnt"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.326017 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-mdhf9"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.327036 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-5q9rs"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.327997 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-machine-config-operator/machine-config-server-rvhq7"] Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.328488 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-rvhq7" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.328616 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.348881 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.363354 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t9zdh\" (UniqueName: \"kubernetes.io/projected/74ab81d8-55e8-47da-adb5-1f616a100a31-kube-api-access-t9zdh\") pod \"downloads-7954f5f757-xdh6w\" (UID: \"74ab81d8-55e8-47da-adb5-1f616a100a31\") " pod="openshift-console/downloads-7954f5f757-xdh6w" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.368799 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.388956 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.408660 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.429099 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.449728 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.476129 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.489338 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.529395 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.549805 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.569522 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.589567 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.608836 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.630403 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.650340 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.671582 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.689937 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.709900 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.730632 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.749537 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.769741 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.789753 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.810260 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.829604 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.849497 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.870196 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.890138 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.909693 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.930235 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.960885 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.969624 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Mar 12 00:10:15 crc kubenswrapper[4948]: I0312 00:10:15.989904 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Mar 12 00:10:16 crc kubenswrapper[4948]: I0312 00:10:16.009973 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Mar 12 00:10:16 crc kubenswrapper[4948]: I0312 00:10:16.030838 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Mar 12 00:10:16 crc kubenswrapper[4948]: I0312 00:10:16.053086 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Mar 12 00:10:16 crc kubenswrapper[4948]: I0312 00:10:16.069503 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Mar 12 00:10:16 crc kubenswrapper[4948]: I0312 00:10:16.089664 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Mar 12 00:10:16 crc kubenswrapper[4948]: I0312 00:10:16.109641 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Mar 12 00:10:16 crc kubenswrapper[4948]: I0312 00:10:16.129191 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Mar 12 00:10:16 crc kubenswrapper[4948]: I0312 00:10:16.149723 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Mar 12 00:10:16 crc kubenswrapper[4948]: I0312 00:10:16.170163 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Mar 12 00:10:16 crc kubenswrapper[4948]: I0312 00:10:16.189622 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Mar 12 00:10:16 crc kubenswrapper[4948]: I0312 00:10:16.208276 4948 request.go:700] Waited for 1.010742079s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-marketplace/secrets?fieldSelector=metadata.name%3Dmarketplace-operator-dockercfg-5nsgg&limit=500&resourceVersion=0 Mar 12 00:10:16 crc kubenswrapper[4948]: I0312 00:10:16.211043 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Mar 12 00:10:16 crc kubenswrapper[4948]: I0312 00:10:16.229761 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Mar 12 00:10:16 crc kubenswrapper[4948]: I0312 00:10:16.249517 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Mar 12 00:10:16 crc kubenswrapper[4948]: I0312 00:10:16.269406 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Mar 12 00:10:16 crc kubenswrapper[4948]: I0312 00:10:16.297238 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Mar 12 00:10:16 crc kubenswrapper[4948]: I0312 00:10:16.309393 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Mar 12 00:10:16 crc kubenswrapper[4948]: I0312 00:10:16.329704 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Mar 12 00:10:16 crc kubenswrapper[4948]: I0312 00:10:16.349827 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Mar 12 00:10:16 crc kubenswrapper[4948]: I0312 00:10:16.369293 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Mar 12 00:10:16 crc kubenswrapper[4948]: I0312 00:10:16.389973 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Mar 12 00:10:16 crc kubenswrapper[4948]: I0312 00:10:16.408583 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Mar 12 00:10:16 crc kubenswrapper[4948]: I0312 00:10:16.429569 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Mar 12 00:10:16 crc kubenswrapper[4948]: I0312 00:10:16.469775 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Mar 12 00:10:16 crc kubenswrapper[4948]: I0312 00:10:16.489916 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Mar 12 00:10:16 crc kubenswrapper[4948]: I0312 00:10:16.510749 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Mar 12 00:10:16 crc kubenswrapper[4948]: I0312 00:10:16.530828 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Mar 12 00:10:16 crc kubenswrapper[4948]: I0312 00:10:16.550108 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Mar 12 00:10:16 crc kubenswrapper[4948]: I0312 00:10:16.570840 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Mar 12 00:10:16 crc kubenswrapper[4948]: I0312 00:10:16.589199 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 12 00:10:16 crc kubenswrapper[4948]: I0312 00:10:16.609593 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 12 00:10:16 crc kubenswrapper[4948]: I0312 00:10:16.630260 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Mar 12 00:10:16 crc kubenswrapper[4948]: I0312 00:10:16.649459 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 00:10:16 crc kubenswrapper[4948]: I0312 00:10:16.669670 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 00:10:16 crc kubenswrapper[4948]: I0312 00:10:16.689754 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Mar 12 00:10:16 crc kubenswrapper[4948]: I0312 00:10:16.709699 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Mar 12 00:10:16 crc kubenswrapper[4948]: I0312 00:10:16.730075 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Mar 12 00:10:16 crc kubenswrapper[4948]: I0312 00:10:16.749961 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Mar 12 00:10:16 crc kubenswrapper[4948]: I0312 00:10:16.769113 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Mar 12 00:10:16 crc kubenswrapper[4948]: I0312 00:10:16.790674 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Mar 12 00:10:16 crc kubenswrapper[4948]: I0312 00:10:16.810629 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Mar 12 00:10:16 crc kubenswrapper[4948]: I0312 00:10:16.830385 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Mar 12 00:10:16 crc kubenswrapper[4948]: I0312 00:10:16.850336 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Mar 12 00:10:16 crc kubenswrapper[4948]: I0312 00:10:16.869932 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Mar 12 00:10:16 crc kubenswrapper[4948]: I0312 00:10:16.889275 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Mar 12 00:10:16 crc kubenswrapper[4948]: I0312 00:10:16.910888 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Mar 12 00:10:16 crc kubenswrapper[4948]: I0312 00:10:16.929724 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Mar 12 00:10:16 crc kubenswrapper[4948]: I0312 00:10:16.958733 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Mar 12 00:10:16 crc kubenswrapper[4948]: I0312 00:10:16.969555 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Mar 12 00:10:16 crc kubenswrapper[4948]: I0312 00:10:16.989265 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.009862 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.053590 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4xvzv\" (UniqueName: \"kubernetes.io/projected/a12c201c-f0ce-4519-8873-e7ec122e7fec-kube-api-access-4xvzv\") pod \"openshift-config-operator-7777fb866f-sg87j\" (UID: \"a12c201c-f0ce-4519-8873-e7ec122e7fec\") " pod="openshift-config-operator/openshift-config-operator-7777fb866f-sg87j" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.057206 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-config-operator/openshift-config-operator-7777fb866f-sg87j" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.078959 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l8vtg\" (UniqueName: \"kubernetes.io/projected/2040ed8d-ae32-4909-9161-bc7ffbe49edd-kube-api-access-l8vtg\") pod \"apiserver-76f77b778f-6jfc4\" (UID: \"2040ed8d-ae32-4909-9161-bc7ffbe49edd\") " pod="openshift-apiserver/apiserver-76f77b778f-6jfc4" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.093075 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dcmgg\" (UniqueName: \"kubernetes.io/projected/a544acdb-b758-4ea8-a43d-8270fc454901-kube-api-access-dcmgg\") pod \"machine-api-operator-5694c8668f-g7g94\" (UID: \"a544acdb-b758-4ea8-a43d-8270fc454901\") " pod="openshift-machine-api/machine-api-operator-5694c8668f-g7g94" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.111824 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmk5h\" (UniqueName: \"kubernetes.io/projected/a6da8984-cc2d-4b58-b6da-d7fc2d1e1725-kube-api-access-fmk5h\") pod \"machine-approver-56656f9798-n4545\" (UID: \"a6da8984-cc2d-4b58-b6da-d7fc2d1e1725\") " pod="openshift-cluster-machine-approver/machine-approver-56656f9798-n4545" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.128166 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gl6qb\" (UniqueName: \"kubernetes.io/projected/5c9bc3de-a78c-44d5-90ca-f57a625e125a-kube-api-access-gl6qb\") pod \"authentication-operator-69f744f599-5cwjt\" (UID: \"5c9bc3de-a78c-44d5-90ca-f57a625e125a\") " pod="openshift-authentication-operator/authentication-operator-69f744f599-5cwjt" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.148949 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6q7g6\" (UniqueName: \"kubernetes.io/projected/60024b0b-f817-477f-83f1-e53cbc1e05b0-kube-api-access-6q7g6\") pod \"route-controller-manager-6576b87f9c-rp8sv\" (UID: \"60024b0b-f817-477f-83f1-e53cbc1e05b0\") " pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rp8sv" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.165718 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-2j6ml\" (UniqueName: \"kubernetes.io/projected/9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5-kube-api-access-2j6ml\") pod \"controller-manager-879f6c89f-9tpt4\" (UID: \"9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5\") " pod="openshift-controller-manager/controller-manager-879f6c89f-9tpt4" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.173007 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/machine-api-operator-5694c8668f-g7g94" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.179054 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dgm4m\" (UniqueName: \"kubernetes.io/projected/f90934d6-9305-4513-8d84-cd6a6fcc4a59-kube-api-access-dgm4m\") pod \"oauth-openshift-558db77b4-qsf8x\" (UID: \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\") " pod="openshift-authentication/oauth-openshift-558db77b4-qsf8x" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.202364 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vrqxj\" (UniqueName: \"kubernetes.io/projected/b36cdace-7dbb-4e8a-b0f0-6e93615612a8-kube-api-access-vrqxj\") pod \"openshift-apiserver-operator-796bbdcf4f-vjnjc\" (UID: \"b36cdace-7dbb-4e8a-b0f0-6e93615612a8\") " pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vjnjc" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.211890 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-9tpt4" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.224522 4948 request.go:700] Waited for 1.904791399s due to client-side throttling, not priority and fairness, request: GET:https://api-int.crc.testing:6443/api/v1/namespaces/openshift-dns/secrets?fieldSelector=metadata.name%3Ddns-dockercfg-jwfmh&limit=500&resourceVersion=0 Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.226239 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.226568 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-v9fgz\" (UniqueName: \"kubernetes.io/projected/f980811d-53fd-4c64-96c5-19283afcbb1f-kube-api-access-v9fgz\") pod \"apiserver-7bbb656c7d-fx7lk\" (UID: \"f980811d-53fd-4c64-96c5-19283afcbb1f\") " pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fx7lk" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.234136 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fx7lk" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.249185 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.262582 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vjnjc" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.265462 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.284980 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.297606 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-config-operator/openshift-config-operator-7777fb866f-sg87j"] Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.305749 4948 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.311198 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rp8sv" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.318954 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication-operator/authentication-operator-69f744f599-5cwjt" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.324996 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.333960 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-n4545" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.338968 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-qsf8x" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.346734 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.346879 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/machine-api-operator-5694c8668f-g7g94"] Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.358514 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-apiserver/apiserver-76f77b778f-6jfc4" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.366275 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Mar 12 00:10:17 crc kubenswrapper[4948]: W0312 00:10:17.368496 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda544acdb_b758_4ea8_a43d_8270fc454901.slice/crio-bcaf812660f16c8387771b7d5ce1db8a404b09f01e854e8133878577c66e81f1 WatchSource:0}: Error finding container bcaf812660f16c8387771b7d5ce1db8a404b09f01e854e8133878577c66e81f1: Status 404 returned error can't find the container with id bcaf812660f16c8387771b7d5ce1db8a404b09f01e854e8133878577c66e81f1 Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.386839 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Mar 12 00:10:17 crc kubenswrapper[4948]: W0312 00:10:17.388105 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda6da8984_cc2d_4b58_b6da_d7fc2d1e1725.slice/crio-6180b9ac487cbd1a227e1d2e55c90b5e78613c37a792d13440fe9f7635826981 WatchSource:0}: Error finding container 6180b9ac487cbd1a227e1d2e55c90b5e78613c37a792d13440fe9f7635826981: Status 404 returned error can't find the container with id 6180b9ac487cbd1a227e1d2e55c90b5e78613c37a792d13440fe9f7635826981 Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.405888 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-9tpt4"] Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.427417 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t9zdh\" (UniqueName: \"kubernetes.io/projected/74ab81d8-55e8-47da-adb5-1f616a100a31-kube-api-access-t9zdh\") pod \"downloads-7954f5f757-xdh6w\" (UID: \"74ab81d8-55e8-47da-adb5-1f616a100a31\") " pod="openshift-console/downloads-7954f5f757-xdh6w" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.493032 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e5a84d1-542e-4b58-b991-611bd3202bcf-config\") pod \"etcd-operator-b45778765-pzgdk\" (UID: \"5e5a84d1-542e-4b58-b991-611bd3202bcf\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pzgdk" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.493059 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2b9e45c7-edbb-4cbb-9ee2-ae7efb61fbb3-bound-sa-token\") pod \"ingress-operator-5b745b69d9-z6v9h\" (UID: \"2b9e45c7-edbb-4cbb-9ee2-ae7efb61fbb3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-z6v9h" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.493081 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/db40ba09-a72f-43d5-8037-e71c96e2bf5b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-zkbhv\" (UID: \"db40ba09-a72f-43d5-8037-e71c96e2bf5b\") " pod="openshift-marketplace/marketplace-operator-79b997595-zkbhv" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.493117 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ngczm\" (UniqueName: \"kubernetes.io/projected/70b489ae-f74b-4e9f-a5b2-71ec1c1f002c-kube-api-access-ngczm\") pod \"dns-operator-744455d44c-xs8b7\" (UID: \"70b489ae-f74b-4e9f-a5b2-71ec1c1f002c\") " pod="openshift-dns-operator/dns-operator-744455d44c-xs8b7" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.493140 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49fe2194-f940-4e0a-a214-fc6080068163-config\") pod \"service-ca-operator-777779d784-kjggj\" (UID: \"49fe2194-f940-4e0a-a214-fc6080068163\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-kjggj" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.493198 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbrg4\" (UniqueName: \"kubernetes.io/projected/2c8b7f32-4f80-4f95-b361-7d4687416711-kube-api-access-lbrg4\") pod \"image-pruner-29554560-slpk9\" (UID: \"2c8b7f32-4f80-4f95-b361-7d4687416711\") " pod="openshift-image-registry/image-pruner-29554560-slpk9" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.493214 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2869c0eb-c1c6-4db7-9cf5-17e3300fa14d-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-4fn9j\" (UID: \"2869c0eb-c1c6-4db7-9cf5-17e3300fa14d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4fn9j" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.493228 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1196fdb1-9c09-4982-9c37-35f46fabc1c5-proxy-tls\") pod \"machine-config-operator-74547568cd-9k744\" (UID: \"1196fdb1-9c09-4982-9c37-35f46fabc1c5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9k744" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.493254 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4dc52af7-648a-40e7-bdae-50672f5143dc-config\") pod \"kube-apiserver-operator-766d6c64bb-gc7vl\" (UID: \"4dc52af7-648a-40e7-bdae-50672f5143dc\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gc7vl" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.493285 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/892017c6-01a1-4798-9d23-fff4f44c6932-console-serving-cert\") pod \"console-f9d7485db-qq8gg\" (UID: \"892017c6-01a1-4798-9d23-fff4f44c6932\") " pod="openshift-console/console-f9d7485db-qq8gg" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.493312 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gfgnq\" (UniqueName: \"kubernetes.io/projected/892017c6-01a1-4798-9d23-fff4f44c6932-kube-api-access-gfgnq\") pod \"console-f9d7485db-qq8gg\" (UID: \"892017c6-01a1-4798-9d23-fff4f44c6932\") " pod="openshift-console/console-f9d7485db-qq8gg" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.493328 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b265c22f-ab92-48e4-9c3e-db6ea6956f6f-config\") pod \"console-operator-58897d9998-z7w95\" (UID: \"b265c22f-ab92-48e4-9c3e-db6ea6956f6f\") " pod="openshift-console-operator/console-operator-58897d9998-z7w95" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.493363 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.493381 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7k6hs\" (UniqueName: \"kubernetes.io/projected/09e21136-b8c7-4680-ab71-73a86766a7b8-kube-api-access-7k6hs\") pod \"multus-admission-controller-857f4d67dd-d24n2\" (UID: \"09e21136-b8c7-4680-ab71-73a86766a7b8\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-d24n2" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.493398 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/892017c6-01a1-4798-9d23-fff4f44c6932-oauth-serving-cert\") pod \"console-f9d7485db-qq8gg\" (UID: \"892017c6-01a1-4798-9d23-fff4f44c6932\") " pod="openshift-console/console-f9d7485db-qq8gg" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.493414 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2b9e45c7-edbb-4cbb-9ee2-ae7efb61fbb3-trusted-ca\") pod \"ingress-operator-5b745b69d9-z6v9h\" (UID: \"2b9e45c7-edbb-4cbb-9ee2-ae7efb61fbb3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-z6v9h" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.493448 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/29cf53b1-0215-4dd4-a19f-24a48d4ded9d-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-rfxxc\" (UID: \"29cf53b1-0215-4dd4-a19f-24a48d4ded9d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rfxxc" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.493462 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a79948db-9342-439c-95ae-128c8bd0524d-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-qc4wh\" (UID: \"a79948db-9342-439c-95ae-128c8bd0524d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qc4wh" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.493477 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/af75ce0a-d6bb-48db-a422-08b1d87e1d3f-installation-pull-secrets\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.493492 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7kscx\" (UniqueName: \"kubernetes.io/projected/49fe2194-f940-4e0a-a214-fc6080068163-kube-api-access-7kscx\") pod \"service-ca-operator-777779d784-kjggj\" (UID: \"49fe2194-f940-4e0a-a214-fc6080068163\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-kjggj" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.493514 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/af75ce0a-d6bb-48db-a422-08b1d87e1d3f-trusted-ca\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.493529 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/af75ce0a-d6bb-48db-a422-08b1d87e1d3f-ca-trust-extracted\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.493545 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/1196fdb1-9c09-4982-9c37-35f46fabc1c5-images\") pod \"machine-config-operator-74547568cd-9k744\" (UID: \"1196fdb1-9c09-4982-9c37-35f46fabc1c5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9k744" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.493558 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/5e5a84d1-542e-4b58-b991-611bd3202bcf-etcd-service-ca\") pod \"etcd-operator-b45778765-pzgdk\" (UID: \"5e5a84d1-542e-4b58-b991-611bd3202bcf\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pzgdk" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.493572 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4q42m\" (UniqueName: \"kubernetes.io/projected/b265c22f-ab92-48e4-9c3e-db6ea6956f6f-kube-api-access-4q42m\") pod \"console-operator-58897d9998-z7w95\" (UID: \"b265c22f-ab92-48e4-9c3e-db6ea6956f6f\") " pod="openshift-console-operator/console-operator-58897d9998-z7w95" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.493586 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2869c0eb-c1c6-4db7-9cf5-17e3300fa14d-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-4fn9j\" (UID: \"2869c0eb-c1c6-4db7-9cf5-17e3300fa14d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4fn9j" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.493610 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d08b0c90-3166-4919-88b3-345f5be3efe7-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6m4k7\" (UID: \"d08b0c90-3166-4919-88b3-345f5be3efe7\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6m4k7" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.493641 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/2c8b7f32-4f80-4f95-b361-7d4687416711-serviceca\") pod \"image-pruner-29554560-slpk9\" (UID: \"2c8b7f32-4f80-4f95-b361-7d4687416711\") " pod="openshift-image-registry/image-pruner-29554560-slpk9" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.493673 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wkr7n\" (UniqueName: \"kubernetes.io/projected/1666a214-546a-4063-a627-e990781797c8-kube-api-access-wkr7n\") pod \"kube-storage-version-migrator-operator-b67b599dd-kvh9w\" (UID: \"1666a214-546a-4063-a627-e990781797c8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kvh9w" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.493695 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1666a214-546a-4063-a627-e990781797c8-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-kvh9w\" (UID: \"1666a214-546a-4063-a627-e990781797c8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kvh9w" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.493709 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/70b489ae-f74b-4e9f-a5b2-71ec1c1f002c-metrics-tls\") pod \"dns-operator-744455d44c-xs8b7\" (UID: \"70b489ae-f74b-4e9f-a5b2-71ec1c1f002c\") " pod="openshift-dns-operator/dns-operator-744455d44c-xs8b7" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.493737 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-r5cbw\" (UniqueName: \"kubernetes.io/projected/677936de-2686-46cf-99e9-cae2355c42b4-kube-api-access-r5cbw\") pod \"migrator-59844c95c7-srr9m\" (UID: \"677936de-2686-46cf-99e9-cae2355c42b4\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-srr9m" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.493752 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/5e5a84d1-542e-4b58-b991-611bd3202bcf-etcd-ca\") pod \"etcd-operator-b45778765-pzgdk\" (UID: \"5e5a84d1-542e-4b58-b991-611bd3202bcf\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pzgdk" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.493804 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/2869c0eb-c1c6-4db7-9cf5-17e3300fa14d-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-4fn9j\" (UID: \"2869c0eb-c1c6-4db7-9cf5-17e3300fa14d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4fn9j" Mar 12 00:10:17 crc kubenswrapper[4948]: E0312 00:10:17.495324 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 00:10:17.995308607 +0000 UTC m=+217.450912345 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g4kcb" (UID: "af75ce0a-d6bb-48db-a422-08b1d87e1d3f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.495386 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/892017c6-01a1-4798-9d23-fff4f44c6932-service-ca\") pod \"console-f9d7485db-qq8gg\" (UID: \"892017c6-01a1-4798-9d23-fff4f44c6932\") " pod="openshift-console/console-f9d7485db-qq8gg" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.495408 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rjspx\" (UniqueName: \"kubernetes.io/projected/1196fdb1-9c09-4982-9c37-35f46fabc1c5-kube-api-access-rjspx\") pod \"machine-config-operator-74547568cd-9k744\" (UID: \"1196fdb1-9c09-4982-9c37-35f46fabc1c5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9k744" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.495426 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-k69zs\" (UniqueName: \"kubernetes.io/projected/2869c0eb-c1c6-4db7-9cf5-17e3300fa14d-kube-api-access-k69zs\") pod \"cluster-image-registry-operator-dc59b4c8b-4fn9j\" (UID: \"2869c0eb-c1c6-4db7-9cf5-17e3300fa14d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4fn9j" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.495442 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-htmnx\" (UniqueName: \"kubernetes.io/projected/5e5a84d1-542e-4b58-b991-611bd3202bcf-kube-api-access-htmnx\") pod \"etcd-operator-b45778765-pzgdk\" (UID: \"5e5a84d1-542e-4b58-b991-611bd3202bcf\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pzgdk" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.495459 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1666a214-546a-4063-a627-e990781797c8-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-kvh9w\" (UID: \"1666a214-546a-4063-a627-e990781797c8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kvh9w" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.495520 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5t7dj\" (UniqueName: \"kubernetes.io/projected/e6184ca8-dc6e-4d34-ad0e-87c7107acb82-kube-api-access-5t7dj\") pod \"router-default-5444994796-7vhkr\" (UID: \"e6184ca8-dc6e-4d34-ad0e-87c7107acb82\") " pod="openshift-ingress/router-default-5444994796-7vhkr" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.495550 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2b9e45c7-edbb-4cbb-9ee2-ae7efb61fbb3-metrics-tls\") pod \"ingress-operator-5b745b69d9-z6v9h\" (UID: \"2b9e45c7-edbb-4cbb-9ee2-ae7efb61fbb3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-z6v9h" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.495577 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jl7d7\" (UniqueName: \"kubernetes.io/projected/af75ce0a-d6bb-48db-a422-08b1d87e1d3f-kube-api-access-jl7d7\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.495592 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5e5a84d1-542e-4b58-b991-611bd3202bcf-serving-cert\") pod \"etcd-operator-b45778765-pzgdk\" (UID: \"5e5a84d1-542e-4b58-b991-611bd3202bcf\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pzgdk" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.495609 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qm88d\" (UniqueName: \"kubernetes.io/projected/2352fbc0-94ff-4b0c-b499-d16b94c279df-kube-api-access-qm88d\") pod \"control-plane-machine-set-operator-78cbb6b69f-bj72v\" (UID: \"2352fbc0-94ff-4b0c-b499-d16b94c279df\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bj72v" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.495635 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/892017c6-01a1-4798-9d23-fff4f44c6932-trusted-ca-bundle\") pod \"console-f9d7485db-qq8gg\" (UID: \"892017c6-01a1-4798-9d23-fff4f44c6932\") " pod="openshift-console/console-f9d7485db-qq8gg" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.495652 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/e6184ca8-dc6e-4d34-ad0e-87c7107acb82-default-certificate\") pod \"router-default-5444994796-7vhkr\" (UID: \"e6184ca8-dc6e-4d34-ad0e-87c7107acb82\") " pod="openshift-ingress/router-default-5444994796-7vhkr" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.495668 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/09e21136-b8c7-4680-ab71-73a86766a7b8-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-d24n2\" (UID: \"09e21136-b8c7-4680-ab71-73a86766a7b8\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-d24n2" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.495685 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/db40ba09-a72f-43d5-8037-e71c96e2bf5b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-zkbhv\" (UID: \"db40ba09-a72f-43d5-8037-e71c96e2bf5b\") " pod="openshift-marketplace/marketplace-operator-79b997595-zkbhv" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.495719 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/af75ce0a-d6bb-48db-a422-08b1d87e1d3f-bound-sa-token\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.495965 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/49fe2194-f940-4e0a-a214-fc6080068163-serving-cert\") pod \"service-ca-operator-777779d784-kjggj\" (UID: \"49fe2194-f940-4e0a-a214-fc6080068163\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-kjggj" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.495986 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d08b0c90-3166-4919-88b3-345f5be3efe7-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6m4k7\" (UID: \"d08b0c90-3166-4919-88b3-345f5be3efe7\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6m4k7" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.496014 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5e5a84d1-542e-4b58-b991-611bd3202bcf-etcd-client\") pod \"etcd-operator-b45778765-pzgdk\" (UID: \"5e5a84d1-542e-4b58-b991-611bd3202bcf\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pzgdk" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.496055 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a79948db-9342-439c-95ae-128c8bd0524d-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-qc4wh\" (UID: \"a79948db-9342-439c-95ae-128c8bd0524d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qc4wh" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.496080 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e6184ca8-dc6e-4d34-ad0e-87c7107acb82-metrics-certs\") pod \"router-default-5444994796-7vhkr\" (UID: \"e6184ca8-dc6e-4d34-ad0e-87c7107acb82\") " pod="openshift-ingress/router-default-5444994796-7vhkr" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.496158 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1196fdb1-9c09-4982-9c37-35f46fabc1c5-auth-proxy-config\") pod \"machine-config-operator-74547568cd-9k744\" (UID: \"1196fdb1-9c09-4982-9c37-35f46fabc1c5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9k744" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.496177 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/2352fbc0-94ff-4b0c-b499-d16b94c279df-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-bj72v\" (UID: \"2352fbc0-94ff-4b0c-b499-d16b94c279df\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bj72v" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.496380 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/af75ce0a-d6bb-48db-a422-08b1d87e1d3f-registry-certificates\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.496415 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g7gzh\" (UniqueName: \"kubernetes.io/projected/db40ba09-a72f-43d5-8037-e71c96e2bf5b-kube-api-access-g7gzh\") pod \"marketplace-operator-79b997595-zkbhv\" (UID: \"db40ba09-a72f-43d5-8037-e71c96e2bf5b\") " pod="openshift-marketplace/marketplace-operator-79b997595-zkbhv" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.496508 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qk6mp\" (UniqueName: \"kubernetes.io/projected/29cf53b1-0215-4dd4-a19f-24a48d4ded9d-kube-api-access-qk6mp\") pod \"cluster-samples-operator-665b6dd947-rfxxc\" (UID: \"29cf53b1-0215-4dd4-a19f-24a48d4ded9d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rfxxc" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.496633 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/892017c6-01a1-4798-9d23-fff4f44c6932-console-config\") pod \"console-f9d7485db-qq8gg\" (UID: \"892017c6-01a1-4798-9d23-fff4f44c6932\") " pod="openshift-console/console-f9d7485db-qq8gg" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.496716 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fj9zg\" (UniqueName: \"kubernetes.io/projected/2b9e45c7-edbb-4cbb-9ee2-ae7efb61fbb3-kube-api-access-fj9zg\") pod \"ingress-operator-5b745b69d9-z6v9h\" (UID: \"2b9e45c7-edbb-4cbb-9ee2-ae7efb61fbb3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-z6v9h" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.497188 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4dc52af7-648a-40e7-bdae-50672f5143dc-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-gc7vl\" (UID: \"4dc52af7-648a-40e7-bdae-50672f5143dc\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gc7vl" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.497224 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b265c22f-ab92-48e4-9c3e-db6ea6956f6f-trusted-ca\") pod \"console-operator-58897d9998-z7w95\" (UID: \"b265c22f-ab92-48e4-9c3e-db6ea6956f6f\") " pod="openshift-console-operator/console-operator-58897d9998-z7w95" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.497283 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b265c22f-ab92-48e4-9c3e-db6ea6956f6f-serving-cert\") pod \"console-operator-58897d9998-z7w95\" (UID: \"b265c22f-ab92-48e4-9c3e-db6ea6956f6f\") " pod="openshift-console-operator/console-operator-58897d9998-z7w95" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.497313 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/e6184ca8-dc6e-4d34-ad0e-87c7107acb82-stats-auth\") pod \"router-default-5444994796-7vhkr\" (UID: \"e6184ca8-dc6e-4d34-ad0e-87c7107acb82\") " pod="openshift-ingress/router-default-5444994796-7vhkr" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.497371 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/892017c6-01a1-4798-9d23-fff4f44c6932-console-oauth-config\") pod \"console-f9d7485db-qq8gg\" (UID: \"892017c6-01a1-4798-9d23-fff4f44c6932\") " pod="openshift-console/console-f9d7485db-qq8gg" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.497413 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d08b0c90-3166-4919-88b3-345f5be3efe7-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6m4k7\" (UID: \"d08b0c90-3166-4919-88b3-345f5be3efe7\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6m4k7" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.497428 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e6184ca8-dc6e-4d34-ad0e-87c7107acb82-service-ca-bundle\") pod \"router-default-5444994796-7vhkr\" (UID: \"e6184ca8-dc6e-4d34-ad0e-87c7107acb82\") " pod="openshift-ingress/router-default-5444994796-7vhkr" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.497444 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qtpms\" (UniqueName: \"kubernetes.io/projected/a79948db-9342-439c-95ae-128c8bd0524d-kube-api-access-qtpms\") pod \"openshift-controller-manager-operator-756b6f6bc6-qc4wh\" (UID: \"a79948db-9342-439c-95ae-128c8bd0524d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qc4wh" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.497460 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/af75ce0a-d6bb-48db-a422-08b1d87e1d3f-registry-tls\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.497475 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4dc52af7-648a-40e7-bdae-50672f5143dc-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-gc7vl\" (UID: \"4dc52af7-648a-40e7-bdae-50672f5143dc\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gc7vl" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.556476 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-rp8sv"] Mar 12 00:10:17 crc kubenswrapper[4948]: W0312 00:10:17.566381 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod60024b0b_f817_477f_83f1_e53cbc1e05b0.slice/crio-4795f512d0d3efb3414bf29827a15719ed2468da83504acccf6d4089a5361df6 WatchSource:0}: Error finding container 4795f512d0d3efb3414bf29827a15719ed2468da83504acccf6d4089a5361df6: Status 404 returned error can't find the container with id 4795f512d0d3efb3414bf29827a15719ed2468da83504acccf6d4089a5361df6 Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.597864 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 00:10:17 crc kubenswrapper[4948]: E0312 00:10:17.597997 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 00:10:18.097973116 +0000 UTC m=+217.553576854 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.598046 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b265c22f-ab92-48e4-9c3e-db6ea6956f6f-serving-cert\") pod \"console-operator-58897d9998-z7w95\" (UID: \"b265c22f-ab92-48e4-9c3e-db6ea6956f6f\") " pod="openshift-console-operator/console-operator-58897d9998-z7w95" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.598069 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/e6184ca8-dc6e-4d34-ad0e-87c7107acb82-stats-auth\") pod \"router-default-5444994796-7vhkr\" (UID: \"e6184ca8-dc6e-4d34-ad0e-87c7107acb82\") " pod="openshift-ingress/router-default-5444994796-7vhkr" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.598087 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qtpms\" (UniqueName: \"kubernetes.io/projected/a79948db-9342-439c-95ae-128c8bd0524d-kube-api-access-qtpms\") pod \"openshift-controller-manager-operator-756b6f6bc6-qc4wh\" (UID: \"a79948db-9342-439c-95ae-128c8bd0524d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qc4wh" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.598109 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/892017c6-01a1-4798-9d23-fff4f44c6932-console-oauth-config\") pod \"console-f9d7485db-qq8gg\" (UID: \"892017c6-01a1-4798-9d23-fff4f44c6932\") " pod="openshift-console/console-f9d7485db-qq8gg" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.598127 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d08b0c90-3166-4919-88b3-345f5be3efe7-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6m4k7\" (UID: \"d08b0c90-3166-4919-88b3-345f5be3efe7\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6m4k7" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.598143 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e6184ca8-dc6e-4d34-ad0e-87c7107acb82-service-ca-bundle\") pod \"router-default-5444994796-7vhkr\" (UID: \"e6184ca8-dc6e-4d34-ad0e-87c7107acb82\") " pod="openshift-ingress/router-default-5444994796-7vhkr" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.598159 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/af75ce0a-d6bb-48db-a422-08b1d87e1d3f-registry-tls\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.598175 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4dc52af7-648a-40e7-bdae-50672f5143dc-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-gc7vl\" (UID: \"4dc52af7-648a-40e7-bdae-50672f5143dc\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gc7vl" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.598195 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7d1dc3d3-060c-4aba-aa54-0a3ea767f299-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-v59sv\" (UID: \"7d1dc3d3-060c-4aba-aa54-0a3ea767f299\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-v59sv" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.598211 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-q4687\" (UniqueName: \"kubernetes.io/projected/3e538a31-98b5-4353-8038-4a7580be89dd-kube-api-access-q4687\") pod \"packageserver-d55dfcdfc-5qjdn\" (UID: \"3e538a31-98b5-4353-8038-4a7580be89dd\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5qjdn" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.598231 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e5a84d1-542e-4b58-b991-611bd3202bcf-config\") pod \"etcd-operator-b45778765-pzgdk\" (UID: \"5e5a84d1-542e-4b58-b991-611bd3202bcf\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pzgdk" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.598227 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication-operator/authentication-operator-69f744f599-5cwjt"] Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.598246 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2b9e45c7-edbb-4cbb-9ee2-ae7efb61fbb3-bound-sa-token\") pod \"ingress-operator-5b745b69d9-z6v9h\" (UID: \"2b9e45c7-edbb-4cbb-9ee2-ae7efb61fbb3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-z6v9h" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.598325 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/db40ba09-a72f-43d5-8037-e71c96e2bf5b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-zkbhv\" (UID: \"db40ba09-a72f-43d5-8037-e71c96e2bf5b\") " pod="openshift-marketplace/marketplace-operator-79b997595-zkbhv" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.598375 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ngczm\" (UniqueName: \"kubernetes.io/projected/70b489ae-f74b-4e9f-a5b2-71ec1c1f002c-kube-api-access-ngczm\") pod \"dns-operator-744455d44c-xs8b7\" (UID: \"70b489ae-f74b-4e9f-a5b2-71ec1c1f002c\") " pod="openshift-dns-operator/dns-operator-744455d44c-xs8b7" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.598396 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/141bb112-1f07-4db5-9a4e-b7d94f12c5e2-signing-key\") pod \"service-ca-9c57cc56f-q79m4\" (UID: \"141bb112-1f07-4db5-9a4e-b7d94f12c5e2\") " pod="openshift-service-ca/service-ca-9c57cc56f-q79m4" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.598417 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49fe2194-f940-4e0a-a214-fc6080068163-config\") pod \"service-ca-operator-777779d784-kjggj\" (UID: \"49fe2194-f940-4e0a-a214-fc6080068163\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-kjggj" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.598435 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/bd97cad6-0d6d-4bec-9d2f-e17b03b220f6-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-lkhdg\" (UID: \"bd97cad6-0d6d-4bec-9d2f-e17b03b220f6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lkhdg" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.598451 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/4ba12474-c71f-429c-bcdf-9d74eb0b4b7a-registration-dir\") pod \"csi-hostpathplugin-5q9rs\" (UID: \"4ba12474-c71f-429c-bcdf-9d74eb0b4b7a\") " pod="hostpath-provisioner/csi-hostpathplugin-5q9rs" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.598465 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rs2ns\" (UniqueName: \"kubernetes.io/projected/4ba12474-c71f-429c-bcdf-9d74eb0b4b7a-kube-api-access-rs2ns\") pod \"csi-hostpathplugin-5q9rs\" (UID: \"4ba12474-c71f-429c-bcdf-9d74eb0b4b7a\") " pod="hostpath-provisioner/csi-hostpathplugin-5q9rs" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.598484 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbrg4\" (UniqueName: \"kubernetes.io/projected/2c8b7f32-4f80-4f95-b361-7d4687416711-kube-api-access-lbrg4\") pod \"image-pruner-29554560-slpk9\" (UID: \"2c8b7f32-4f80-4f95-b361-7d4687416711\") " pod="openshift-image-registry/image-pruner-29554560-slpk9" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.598499 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-hsk7b\" (UniqueName: \"kubernetes.io/projected/15828966-e769-43d4-b4a3-7a509fa72814-kube-api-access-hsk7b\") pod \"ingress-canary-mdhf9\" (UID: \"15828966-e769-43d4-b4a3-7a509fa72814\") " pod="openshift-ingress-canary/ingress-canary-mdhf9" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.598521 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2869c0eb-c1c6-4db7-9cf5-17e3300fa14d-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-4fn9j\" (UID: \"2869c0eb-c1c6-4db7-9cf5-17e3300fa14d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4fn9j" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.598535 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1196fdb1-9c09-4982-9c37-35f46fabc1c5-proxy-tls\") pod \"machine-config-operator-74547568cd-9k744\" (UID: \"1196fdb1-9c09-4982-9c37-35f46fabc1c5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9k744" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.599267 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4dc52af7-648a-40e7-bdae-50672f5143dc-config\") pod \"kube-apiserver-operator-766d6c64bb-gc7vl\" (UID: \"4dc52af7-648a-40e7-bdae-50672f5143dc\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gc7vl" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.599469 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/892017c6-01a1-4798-9d23-fff4f44c6932-console-serving-cert\") pod \"console-f9d7485db-qq8gg\" (UID: \"892017c6-01a1-4798-9d23-fff4f44c6932\") " pod="openshift-console/console-f9d7485db-qq8gg" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.601242 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/db40ba09-a72f-43d5-8037-e71c96e2bf5b-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-zkbhv\" (UID: \"db40ba09-a72f-43d5-8037-e71c96e2bf5b\") " pod="openshift-marketplace/marketplace-operator-79b997595-zkbhv" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.604234 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/e6184ca8-dc6e-4d34-ad0e-87c7107acb82-service-ca-bundle\") pod \"router-default-5444994796-7vhkr\" (UID: \"e6184ca8-dc6e-4d34-ad0e-87c7107acb82\") " pod="openshift-ingress/router-default-5444994796-7vhkr" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.604607 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/d08b0c90-3166-4919-88b3-345f5be3efe7-config\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6m4k7\" (UID: \"d08b0c90-3166-4919-88b3-345f5be3efe7\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6m4k7" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.605154 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/49fe2194-f940-4e0a-a214-fc6080068163-config\") pod \"service-ca-operator-777779d784-kjggj\" (UID: \"49fe2194-f940-4e0a-a214-fc6080068163\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-kjggj" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.605266 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/4dc52af7-648a-40e7-bdae-50672f5143dc-config\") pod \"kube-apiserver-operator-766d6c64bb-gc7vl\" (UID: \"4dc52af7-648a-40e7-bdae-50672f5143dc\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gc7vl" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.605441 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gfgnq\" (UniqueName: \"kubernetes.io/projected/892017c6-01a1-4798-9d23-fff4f44c6932-kube-api-access-gfgnq\") pod \"console-f9d7485db-qq8gg\" (UID: \"892017c6-01a1-4798-9d23-fff4f44c6932\") " pod="openshift-console/console-f9d7485db-qq8gg" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.605847 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-serving-cert\" (UniqueName: \"kubernetes.io/secret/892017c6-01a1-4798-9d23-fff4f44c6932-console-serving-cert\") pod \"console-f9d7485db-qq8gg\" (UID: \"892017c6-01a1-4798-9d23-fff4f44c6932\") " pod="openshift-console/console-f9d7485db-qq8gg" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.605927 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-oauth-config\" (UniqueName: \"kubernetes.io/secret/892017c6-01a1-4798-9d23-fff4f44c6932-console-oauth-config\") pod \"console-f9d7485db-qq8gg\" (UID: \"892017c6-01a1-4798-9d23-fff4f44c6932\") " pod="openshift-console/console-f9d7485db-qq8gg" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.605945 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b265c22f-ab92-48e4-9c3e-db6ea6956f6f-config\") pod \"console-operator-58897d9998-z7w95\" (UID: \"b265c22f-ab92-48e4-9c3e-db6ea6956f6f\") " pod="openshift-console-operator/console-operator-58897d9998-z7w95" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.606717 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.606770 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/5e5a84d1-542e-4b58-b991-611bd3202bcf-config\") pod \"etcd-operator-b45778765-pzgdk\" (UID: \"5e5a84d1-542e-4b58-b991-611bd3202bcf\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pzgdk" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.606792 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/892017c6-01a1-4798-9d23-fff4f44c6932-oauth-serving-cert\") pod \"console-f9d7485db-qq8gg\" (UID: \"892017c6-01a1-4798-9d23-fff4f44c6932\") " pod="openshift-console/console-f9d7485db-qq8gg" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.606816 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2b9e45c7-edbb-4cbb-9ee2-ae7efb61fbb3-trusted-ca\") pod \"ingress-operator-5b745b69d9-z6v9h\" (UID: \"2b9e45c7-edbb-4cbb-9ee2-ae7efb61fbb3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-z6v9h" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.606838 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7k6hs\" (UniqueName: \"kubernetes.io/projected/09e21136-b8c7-4680-ab71-73a86766a7b8-kube-api-access-7k6hs\") pod \"multus-admission-controller-857f4d67dd-d24n2\" (UID: \"09e21136-b8c7-4680-ab71-73a86766a7b8\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-d24n2" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.606868 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/29cf53b1-0215-4dd4-a19f-24a48d4ded9d-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-rfxxc\" (UID: \"29cf53b1-0215-4dd4-a19f-24a48d4ded9d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rfxxc" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.606889 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a79948db-9342-439c-95ae-128c8bd0524d-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-qc4wh\" (UID: \"a79948db-9342-439c-95ae-128c8bd0524d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qc4wh" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.606911 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/af75ce0a-d6bb-48db-a422-08b1d87e1d3f-installation-pull-secrets\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.606930 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7kscx\" (UniqueName: \"kubernetes.io/projected/49fe2194-f940-4e0a-a214-fc6080068163-kube-api-access-7kscx\") pod \"service-ca-operator-777779d784-kjggj\" (UID: \"49fe2194-f940-4e0a-a214-fc6080068163\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-kjggj" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.606953 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/af75ce0a-d6bb-48db-a422-08b1d87e1d3f-trusted-ca\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.606978 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xj5p2\" (UniqueName: \"kubernetes.io/projected/bd97cad6-0d6d-4bec-9d2f-e17b03b220f6-kube-api-access-xj5p2\") pod \"machine-config-controller-84d6567774-lkhdg\" (UID: \"bd97cad6-0d6d-4bec-9d2f-e17b03b220f6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lkhdg" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.607002 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/af75ce0a-d6bb-48db-a422-08b1d87e1d3f-ca-trust-extracted\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.607026 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"images\" (UniqueName: \"kubernetes.io/configmap/1196fdb1-9c09-4982-9c37-35f46fabc1c5-images\") pod \"machine-config-operator-74547568cd-9k744\" (UID: \"1196fdb1-9c09-4982-9c37-35f46fabc1c5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9k744" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.607044 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2869c0eb-c1c6-4db7-9cf5-17e3300fa14d-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-4fn9j\" (UID: \"2869c0eb-c1c6-4db7-9cf5-17e3300fa14d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4fn9j" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.607065 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/5e5a84d1-542e-4b58-b991-611bd3202bcf-etcd-service-ca\") pod \"etcd-operator-b45778765-pzgdk\" (UID: \"5e5a84d1-542e-4b58-b991-611bd3202bcf\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pzgdk" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.607084 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4q42m\" (UniqueName: \"kubernetes.io/projected/b265c22f-ab92-48e4-9c3e-db6ea6956f6f-kube-api-access-4q42m\") pod \"console-operator-58897d9998-z7w95\" (UID: \"b265c22f-ab92-48e4-9c3e-db6ea6956f6f\") " pod="openshift-console-operator/console-operator-58897d9998-z7w95" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.607120 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sjmkx\" (UniqueName: \"kubernetes.io/projected/12c99e25-86bd-4923-ab44-55e571661c0f-kube-api-access-sjmkx\") pod \"olm-operator-6b444d44fb-h6z8v\" (UID: \"12c99e25-86bd-4923-ab44-55e571661c0f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h6z8v" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.607143 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d08b0c90-3166-4919-88b3-345f5be3efe7-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6m4k7\" (UID: \"d08b0c90-3166-4919-88b3-345f5be3efe7\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6m4k7" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.607171 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/4ba12474-c71f-429c-bcdf-9d74eb0b4b7a-csi-data-dir\") pod \"csi-hostpathplugin-5q9rs\" (UID: \"4ba12474-c71f-429c-bcdf-9d74eb0b4b7a\") " pod="hostpath-provisioner/csi-hostpathplugin-5q9rs" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.607199 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/2c8b7f32-4f80-4f95-b361-7d4687416711-serviceca\") pod \"image-pruner-29554560-slpk9\" (UID: \"2c8b7f32-4f80-4f95-b361-7d4687416711\") " pod="openshift-image-registry/image-pruner-29554560-slpk9" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.607220 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wkr7n\" (UniqueName: \"kubernetes.io/projected/1666a214-546a-4063-a627-e990781797c8-kube-api-access-wkr7n\") pod \"kube-storage-version-migrator-operator-b67b599dd-kvh9w\" (UID: \"1666a214-546a-4063-a627-e990781797c8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kvh9w" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.607238 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/d33d5db8-961e-4c7a-950e-cb3f88a3f9b0-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-jlrln\" (UID: \"d33d5db8-961e-4c7a-950e-cb3f88a3f9b0\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jlrln" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.607276 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1666a214-546a-4063-a627-e990781797c8-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-kvh9w\" (UID: \"1666a214-546a-4063-a627-e990781797c8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kvh9w" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.607334 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/70b489ae-f74b-4e9f-a5b2-71ec1c1f002c-metrics-tls\") pod \"dns-operator-744455d44c-xs8b7\" (UID: \"70b489ae-f74b-4e9f-a5b2-71ec1c1f002c\") " pod="openshift-dns-operator/dns-operator-744455d44c-xs8b7" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.607355 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7d1dc3d3-060c-4aba-aa54-0a3ea767f299-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-v59sv\" (UID: \"7d1dc3d3-060c-4aba-aa54-0a3ea767f299\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-v59sv" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.607385 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-r5cbw\" (UniqueName: \"kubernetes.io/projected/677936de-2686-46cf-99e9-cae2355c42b4-kube-api-access-r5cbw\") pod \"migrator-59844c95c7-srr9m\" (UID: \"677936de-2686-46cf-99e9-cae2355c42b4\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-srr9m" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.607420 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/7599787e-ea15-40ca-b365-fdf406e3c337-srv-cert\") pod \"catalog-operator-68c6474976-dc6wb\" (UID: \"7599787e-ea15-40ca-b365-fdf406e3c337\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dc6wb" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.607446 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/5e5a84d1-542e-4b58-b991-611bd3202bcf-etcd-ca\") pod \"etcd-operator-b45778765-pzgdk\" (UID: \"5e5a84d1-542e-4b58-b991-611bd3202bcf\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pzgdk" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.607465 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d4x4t\" (UniqueName: \"kubernetes.io/projected/2851d901-f861-432b-a61e-5e0391f1b2b0-kube-api-access-d4x4t\") pod \"dns-default-prqnt\" (UID: \"2851d901-f861-432b-a61e-5e0391f1b2b0\") " pod="openshift-dns/dns-default-prqnt" Mar 12 00:10:17 crc kubenswrapper[4948]: E0312 00:10:17.607711 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 00:10:18.107689004 +0000 UTC m=+217.563292742 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g4kcb" (UID: "af75ce0a-d6bb-48db-a422-08b1d87e1d3f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.608558 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/b265c22f-ab92-48e4-9c3e-db6ea6956f6f-config\") pod \"console-operator-58897d9998-z7w95\" (UID: \"b265c22f-ab92-48e4-9c3e-db6ea6956f6f\") " pod="openshift-console-operator/console-operator-58897d9998-z7w95" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.609740 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"oauth-serving-cert\" (UniqueName: \"kubernetes.io/configmap/892017c6-01a1-4798-9d23-fff4f44c6932-oauth-serving-cert\") pod \"console-f9d7485db-qq8gg\" (UID: \"892017c6-01a1-4798-9d23-fff4f44c6932\") " pod="openshift-console/console-f9d7485db-qq8gg" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.610845 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/af75ce0a-d6bb-48db-a422-08b1d87e1d3f-trusted-ca\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.610950 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/b265c22f-ab92-48e4-9c3e-db6ea6956f6f-serving-cert\") pod \"console-operator-58897d9998-z7w95\" (UID: \"b265c22f-ab92-48e4-9c3e-db6ea6956f6f\") " pod="openshift-console-operator/console-operator-58897d9998-z7w95" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.611322 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/af75ce0a-d6bb-48db-a422-08b1d87e1d3f-ca-trust-extracted\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.612011 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5b801839-e39a-48a9-9eee-1bb95df5fbfb-config-volume\") pod \"collect-profiles-29554560-8mgv9\" (UID: \"5b801839-e39a-48a9-9eee-1bb95df5fbfb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29554560-8mgv9" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.612601 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/4ba12474-c71f-429c-bcdf-9d74eb0b4b7a-mountpoint-dir\") pod \"csi-hostpathplugin-5q9rs\" (UID: \"4ba12474-c71f-429c-bcdf-9d74eb0b4b7a\") " pod="hostpath-provisioner/csi-hostpathplugin-5q9rs" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.612699 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/2869c0eb-c1c6-4db7-9cf5-17e3300fa14d-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-4fn9j\" (UID: \"2869c0eb-c1c6-4db7-9cf5-17e3300fa14d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4fn9j" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.613424 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2b9e45c7-edbb-4cbb-9ee2-ae7efb61fbb3-trusted-ca\") pod \"ingress-operator-5b745b69d9-z6v9h\" (UID: \"2b9e45c7-edbb-4cbb-9ee2-ae7efb61fbb3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-z6v9h" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.612244 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/1196fdb1-9c09-4982-9c37-35f46fabc1c5-proxy-tls\") pod \"machine-config-operator-74547568cd-9k744\" (UID: \"1196fdb1-9c09-4982-9c37-35f46fabc1c5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9k744" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.612419 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"images\" (UniqueName: \"kubernetes.io/configmap/1196fdb1-9c09-4982-9c37-35f46fabc1c5-images\") pod \"machine-config-operator-74547568cd-9k744\" (UID: \"1196fdb1-9c09-4982-9c37-35f46fabc1c5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9k744" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.613869 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d1dc3d3-060c-4aba-aa54-0a3ea767f299-config\") pod \"kube-controller-manager-operator-78b949d7b-v59sv\" (UID: \"7d1dc3d3-060c-4aba-aa54-0a3ea767f299\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-v59sv" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.613921 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3e538a31-98b5-4353-8038-4a7580be89dd-webhook-cert\") pod \"packageserver-d55dfcdfc-5qjdn\" (UID: \"3e538a31-98b5-4353-8038-4a7580be89dd\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5qjdn" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.613951 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-service-ca\" (UniqueName: \"kubernetes.io/configmap/5e5a84d1-542e-4b58-b991-611bd3202bcf-etcd-service-ca\") pod \"etcd-operator-b45778765-pzgdk\" (UID: \"5e5a84d1-542e-4b58-b991-611bd3202bcf\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pzgdk" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.614106 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-qsf8x"] Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.613939 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/cadf90b6-df8c-48fa-b878-68f2c7d7e49f-node-bootstrap-token\") pod \"machine-config-server-rvhq7\" (UID: \"cadf90b6-df8c-48fa-b878-68f2c7d7e49f\") " pod="openshift-machine-config-operator/machine-config-server-rvhq7" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.614170 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/892017c6-01a1-4798-9d23-fff4f44c6932-service-ca\") pod \"console-f9d7485db-qq8gg\" (UID: \"892017c6-01a1-4798-9d23-fff4f44c6932\") " pod="openshift-console/console-f9d7485db-qq8gg" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.614200 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rjspx\" (UniqueName: \"kubernetes.io/projected/1196fdb1-9c09-4982-9c37-35f46fabc1c5-kube-api-access-rjspx\") pod \"machine-config-operator-74547568cd-9k744\" (UID: \"1196fdb1-9c09-4982-9c37-35f46fabc1c5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9k744" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.614239 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/bd97cad6-0d6d-4bec-9d2f-e17b03b220f6-proxy-tls\") pod \"machine-config-controller-84d6567774-lkhdg\" (UID: \"bd97cad6-0d6d-4bec-9d2f-e17b03b220f6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lkhdg" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.615403 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"service-ca\" (UniqueName: \"kubernetes.io/configmap/892017c6-01a1-4798-9d23-fff4f44c6932-service-ca\") pod \"console-f9d7485db-qq8gg\" (UID: \"892017c6-01a1-4798-9d23-fff4f44c6932\") " pod="openshift-console/console-f9d7485db-qq8gg" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.615336 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/2c8b7f32-4f80-4f95-b361-7d4687416711-serviceca\") pod \"image-pruner-29554560-slpk9\" (UID: \"2c8b7f32-4f80-4f95-b361-7d4687416711\") " pod="openshift-image-registry/image-pruner-29554560-slpk9" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.616853 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/af75ce0a-d6bb-48db-a422-08b1d87e1d3f-registry-tls\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.617767 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"samples-operator-tls\" (UniqueName: \"kubernetes.io/secret/29cf53b1-0215-4dd4-a19f-24a48d4ded9d-samples-operator-tls\") pod \"cluster-samples-operator-665b6dd947-rfxxc\" (UID: \"29cf53b1-0215-4dd4-a19f-24a48d4ded9d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rfxxc" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.617824 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-k69zs\" (UniqueName: \"kubernetes.io/projected/2869c0eb-c1c6-4db7-9cf5-17e3300fa14d-kube-api-access-k69zs\") pod \"cluster-image-registry-operator-dc59b4c8b-4fn9j\" (UID: \"2869c0eb-c1c6-4db7-9cf5-17e3300fa14d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4fn9j" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.617854 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-htmnx\" (UniqueName: \"kubernetes.io/projected/5e5a84d1-542e-4b58-b991-611bd3202bcf-kube-api-access-htmnx\") pod \"etcd-operator-b45778765-pzgdk\" (UID: \"5e5a84d1-542e-4b58-b991-611bd3202bcf\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pzgdk" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.617878 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1666a214-546a-4063-a627-e990781797c8-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-kvh9w\" (UID: \"1666a214-546a-4063-a627-e990781797c8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kvh9w" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.618238 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/a79948db-9342-439c-95ae-128c8bd0524d-serving-cert\") pod \"openshift-controller-manager-operator-756b6f6bc6-qc4wh\" (UID: \"a79948db-9342-439c-95ae-128c8bd0524d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qc4wh" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.618349 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/1666a214-546a-4063-a627-e990781797c8-config\") pod \"kube-storage-version-migrator-operator-b67b599dd-kvh9w\" (UID: \"1666a214-546a-4063-a627-e990781797c8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kvh9w" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.618668 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/70b489ae-f74b-4e9f-a5b2-71ec1c1f002c-metrics-tls\") pod \"dns-operator-744455d44c-xs8b7\" (UID: \"70b489ae-f74b-4e9f-a5b2-71ec1c1f002c\") " pod="openshift-dns-operator/dns-operator-744455d44c-xs8b7" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.618796 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5t7dj\" (UniqueName: \"kubernetes.io/projected/e6184ca8-dc6e-4d34-ad0e-87c7107acb82-kube-api-access-5t7dj\") pod \"router-default-5444994796-7vhkr\" (UID: \"e6184ca8-dc6e-4d34-ad0e-87c7107acb82\") " pod="openshift-ingress/router-default-5444994796-7vhkr" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.618845 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2b9e45c7-edbb-4cbb-9ee2-ae7efb61fbb3-metrics-tls\") pod \"ingress-operator-5b745b69d9-z6v9h\" (UID: \"2b9e45c7-edbb-4cbb-9ee2-ae7efb61fbb3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-z6v9h" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.619101 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5e5a84d1-542e-4b58-b991-611bd3202bcf-serving-cert\") pod \"etcd-operator-b45778765-pzgdk\" (UID: \"5e5a84d1-542e-4b58-b991-611bd3202bcf\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pzgdk" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.619482 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-ca\" (UniqueName: \"kubernetes.io/configmap/5e5a84d1-542e-4b58-b991-611bd3202bcf-etcd-ca\") pod \"etcd-operator-b45778765-pzgdk\" (UID: \"5e5a84d1-542e-4b58-b991-611bd3202bcf\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pzgdk" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.619847 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/af75ce0a-d6bb-48db-a422-08b1d87e1d3f-installation-pull-secrets\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.620013 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jl7d7\" (UniqueName: \"kubernetes.io/projected/af75ce0a-d6bb-48db-a422-08b1d87e1d3f-kube-api-access-jl7d7\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.620246 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qm88d\" (UniqueName: \"kubernetes.io/projected/2352fbc0-94ff-4b0c-b499-d16b94c279df-kube-api-access-qm88d\") pod \"control-plane-machine-set-operator-78cbb6b69f-bj72v\" (UID: \"2352fbc0-94ff-4b0c-b499-d16b94c279df\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bj72v" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.620281 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/892017c6-01a1-4798-9d23-fff4f44c6932-trusted-ca-bundle\") pod \"console-f9d7485db-qq8gg\" (UID: \"892017c6-01a1-4798-9d23-fff4f44c6932\") " pod="openshift-console/console-f9d7485db-qq8gg" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.620311 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/e6184ca8-dc6e-4d34-ad0e-87c7107acb82-default-certificate\") pod \"router-default-5444994796-7vhkr\" (UID: \"e6184ca8-dc6e-4d34-ad0e-87c7107acb82\") " pod="openshift-ingress/router-default-5444994796-7vhkr" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.620334 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2851d901-f861-432b-a61e-5e0391f1b2b0-metrics-tls\") pod \"dns-default-prqnt\" (UID: \"2851d901-f861-432b-a61e-5e0391f1b2b0\") " pod="openshift-dns/dns-default-prqnt" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.620358 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/7599787e-ea15-40ca-b365-fdf406e3c337-profile-collector-cert\") pod \"catalog-operator-68c6474976-dc6wb\" (UID: \"7599787e-ea15-40ca-b365-fdf406e3c337\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dc6wb" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.620388 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/141bb112-1f07-4db5-9a4e-b7d94f12c5e2-signing-cabundle\") pod \"service-ca-9c57cc56f-q79m4\" (UID: \"141bb112-1f07-4db5-9a4e-b7d94f12c5e2\") " pod="openshift-service-ca/service-ca-9c57cc56f-q79m4" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.620417 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/af75ce0a-d6bb-48db-a422-08b1d87e1d3f-bound-sa-token\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.620437 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/09e21136-b8c7-4680-ab71-73a86766a7b8-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-d24n2\" (UID: \"09e21136-b8c7-4680-ab71-73a86766a7b8\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-d24n2" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.620463 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/db40ba09-a72f-43d5-8037-e71c96e2bf5b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-zkbhv\" (UID: \"db40ba09-a72f-43d5-8037-e71c96e2bf5b\") " pod="openshift-marketplace/marketplace-operator-79b997595-zkbhv" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.620485 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/4ba12474-c71f-429c-bcdf-9d74eb0b4b7a-socket-dir\") pod \"csi-hostpathplugin-5q9rs\" (UID: \"4ba12474-c71f-429c-bcdf-9d74eb0b4b7a\") " pod="hostpath-provisioner/csi-hostpathplugin-5q9rs" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.620505 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/12c99e25-86bd-4923-ab44-55e571661c0f-srv-cert\") pod \"olm-operator-6b444d44fb-h6z8v\" (UID: \"12c99e25-86bd-4923-ab44-55e571661c0f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h6z8v" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.620524 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5e5a84d1-542e-4b58-b991-611bd3202bcf-etcd-client\") pod \"etcd-operator-b45778765-pzgdk\" (UID: \"5e5a84d1-542e-4b58-b991-611bd3202bcf\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pzgdk" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.620543 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/49fe2194-f940-4e0a-a214-fc6080068163-serving-cert\") pod \"service-ca-operator-777779d784-kjggj\" (UID: \"49fe2194-f940-4e0a-a214-fc6080068163\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-kjggj" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.620567 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d08b0c90-3166-4919-88b3-345f5be3efe7-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6m4k7\" (UID: \"d08b0c90-3166-4919-88b3-345f5be3efe7\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6m4k7" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.620594 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/cadf90b6-df8c-48fa-b878-68f2c7d7e49f-certs\") pod \"machine-config-server-rvhq7\" (UID: \"cadf90b6-df8c-48fa-b878-68f2c7d7e49f\") " pod="openshift-machine-config-operator/machine-config-server-rvhq7" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.620616 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a79948db-9342-439c-95ae-128c8bd0524d-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-qc4wh\" (UID: \"a79948db-9342-439c-95ae-128c8bd0524d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qc4wh" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.620634 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kn2xg\" (UniqueName: \"kubernetes.io/projected/d33d5db8-961e-4c7a-950e-cb3f88a3f9b0-kube-api-access-kn2xg\") pod \"package-server-manager-789f6589d5-jlrln\" (UID: \"d33d5db8-961e-4c7a-950e-cb3f88a3f9b0\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jlrln" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.620652 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/15828966-e769-43d4-b4a3-7a509fa72814-cert\") pod \"ingress-canary-mdhf9\" (UID: \"15828966-e769-43d4-b4a3-7a509fa72814\") " pod="openshift-ingress-canary/ingress-canary-mdhf9" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.620672 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/12c99e25-86bd-4923-ab44-55e571661c0f-profile-collector-cert\") pod \"olm-operator-6b444d44fb-h6z8v\" (UID: \"12c99e25-86bd-4923-ab44-55e571661c0f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h6z8v" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.620695 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e6184ca8-dc6e-4d34-ad0e-87c7107acb82-metrics-certs\") pod \"router-default-5444994796-7vhkr\" (UID: \"e6184ca8-dc6e-4d34-ad0e-87c7107acb82\") " pod="openshift-ingress/router-default-5444994796-7vhkr" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.620715 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5b801839-e39a-48a9-9eee-1bb95df5fbfb-secret-volume\") pod \"collect-profiles-29554560-8mgv9\" (UID: \"5b801839-e39a-48a9-9eee-1bb95df5fbfb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29554560-8mgv9" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.620734 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/4ba12474-c71f-429c-bcdf-9d74eb0b4b7a-plugins-dir\") pod \"csi-hostpathplugin-5q9rs\" (UID: \"4ba12474-c71f-429c-bcdf-9d74eb0b4b7a\") " pod="hostpath-provisioner/csi-hostpathplugin-5q9rs" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.620757 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1196fdb1-9c09-4982-9c37-35f46fabc1c5-auth-proxy-config\") pod \"machine-config-operator-74547568cd-9k744\" (UID: \"1196fdb1-9c09-4982-9c37-35f46fabc1c5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9k744" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.620782 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/2352fbc0-94ff-4b0c-b499-d16b94c279df-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-bj72v\" (UID: \"2352fbc0-94ff-4b0c-b499-d16b94c279df\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bj72v" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.620805 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6dvsl\" (UniqueName: \"kubernetes.io/projected/2fb324d6-44d4-440d-ae60-a5356a3a057f-kube-api-access-6dvsl\") pod \"auto-csr-approver-29554570-2b2hb\" (UID: \"2fb324d6-44d4-440d-ae60-a5356a3a057f\") " pod="openshift-infra/auto-csr-approver-29554570-2b2hb" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.620828 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/af75ce0a-d6bb-48db-a422-08b1d87e1d3f-registry-certificates\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.620871 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g7gzh\" (UniqueName: \"kubernetes.io/projected/db40ba09-a72f-43d5-8037-e71c96e2bf5b-kube-api-access-g7gzh\") pod \"marketplace-operator-79b997595-zkbhv\" (UID: \"db40ba09-a72f-43d5-8037-e71c96e2bf5b\") " pod="openshift-marketplace/marketplace-operator-79b997595-zkbhv" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.620892 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-bm4bs\" (UniqueName: \"kubernetes.io/projected/7599787e-ea15-40ca-b365-fdf406e3c337-kube-api-access-bm4bs\") pod \"catalog-operator-68c6474976-dc6wb\" (UID: \"7599787e-ea15-40ca-b365-fdf406e3c337\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dc6wb" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.620916 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qk6mp\" (UniqueName: \"kubernetes.io/projected/29cf53b1-0215-4dd4-a19f-24a48d4ded9d-kube-api-access-qk6mp\") pod \"cluster-samples-operator-665b6dd947-rfxxc\" (UID: \"29cf53b1-0215-4dd4-a19f-24a48d4ded9d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rfxxc" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.620940 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8btlp\" (UniqueName: \"kubernetes.io/projected/cadf90b6-df8c-48fa-b878-68f2c7d7e49f-kube-api-access-8btlp\") pod \"machine-config-server-rvhq7\" (UID: \"cadf90b6-df8c-48fa-b878-68f2c7d7e49f\") " pod="openshift-machine-config-operator/machine-config-server-rvhq7" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.620970 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/892017c6-01a1-4798-9d23-fff4f44c6932-console-config\") pod \"console-f9d7485db-qq8gg\" (UID: \"892017c6-01a1-4798-9d23-fff4f44c6932\") " pod="openshift-console/console-f9d7485db-qq8gg" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.620998 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2851d901-f861-432b-a61e-5e0391f1b2b0-config-volume\") pod \"dns-default-prqnt\" (UID: \"2851d901-f861-432b-a61e-5e0391f1b2b0\") " pod="openshift-dns/dns-default-prqnt" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.621039 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fj9zg\" (UniqueName: \"kubernetes.io/projected/2b9e45c7-edbb-4cbb-9ee2-ae7efb61fbb3-kube-api-access-fj9zg\") pod \"ingress-operator-5b745b69d9-z6v9h\" (UID: \"2b9e45c7-edbb-4cbb-9ee2-ae7efb61fbb3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-z6v9h" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.621065 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3e538a31-98b5-4353-8038-4a7580be89dd-apiservice-cert\") pod \"packageserver-d55dfcdfc-5qjdn\" (UID: \"3e538a31-98b5-4353-8038-4a7580be89dd\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5qjdn" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.621091 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4dc52af7-648a-40e7-bdae-50672f5143dc-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-gc7vl\" (UID: \"4dc52af7-648a-40e7-bdae-50672f5143dc\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gc7vl" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.621107 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/3e538a31-98b5-4353-8038-4a7580be89dd-tmpfs\") pod \"packageserver-d55dfcdfc-5qjdn\" (UID: \"3e538a31-98b5-4353-8038-4a7580be89dd\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5qjdn" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.621148 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fzbkp\" (UniqueName: \"kubernetes.io/projected/5b801839-e39a-48a9-9eee-1bb95df5fbfb-kube-api-access-fzbkp\") pod \"collect-profiles-29554560-8mgv9\" (UID: \"5b801839-e39a-48a9-9eee-1bb95df5fbfb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29554560-8mgv9" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.621171 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b265c22f-ab92-48e4-9c3e-db6ea6956f6f-trusted-ca\") pod \"console-operator-58897d9998-z7w95\" (UID: \"b265c22f-ab92-48e4-9c3e-db6ea6956f6f\") " pod="openshift-console-operator/console-operator-58897d9998-z7w95" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.621191 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-9vs6g\" (UniqueName: \"kubernetes.io/projected/141bb112-1f07-4db5-9a4e-b7d94f12c5e2-kube-api-access-9vs6g\") pod \"service-ca-9c57cc56f-q79m4\" (UID: \"141bb112-1f07-4db5-9a4e-b7d94f12c5e2\") " pod="openshift-service-ca/service-ca-9c57cc56f-q79m4" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.622141 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/1666a214-546a-4063-a627-e990781797c8-serving-cert\") pod \"kube-storage-version-migrator-operator-b67b599dd-kvh9w\" (UID: \"1666a214-546a-4063-a627-e990781797c8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kvh9w" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.622799 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/892017c6-01a1-4798-9d23-fff4f44c6932-trusted-ca-bundle\") pod \"console-f9d7485db-qq8gg\" (UID: \"892017c6-01a1-4798-9d23-fff4f44c6932\") " pod="openshift-console/console-f9d7485db-qq8gg" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.623936 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/1196fdb1-9c09-4982-9c37-35f46fabc1c5-auth-proxy-config\") pod \"machine-config-operator-74547568cd-9k744\" (UID: \"1196fdb1-9c09-4982-9c37-35f46fabc1c5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9k744" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.624231 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/2869c0eb-c1c6-4db7-9cf5-17e3300fa14d-trusted-ca\") pod \"cluster-image-registry-operator-dc59b4c8b-4fn9j\" (UID: \"2869c0eb-c1c6-4db7-9cf5-17e3300fa14d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4fn9j" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.623469 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"console-config\" (UniqueName: \"kubernetes.io/configmap/892017c6-01a1-4798-9d23-fff4f44c6932-console-config\") pod \"console-f9d7485db-qq8gg\" (UID: \"892017c6-01a1-4798-9d23-fff4f44c6932\") " pod="openshift-console/console-f9d7485db-qq8gg" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.625087 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/af75ce0a-d6bb-48db-a422-08b1d87e1d3f-registry-certificates\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.626161 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/a79948db-9342-439c-95ae-128c8bd0524d-config\") pod \"openshift-controller-manager-operator-756b6f6bc6-qc4wh\" (UID: \"a79948db-9342-439c-95ae-128c8bd0524d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qc4wh" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.626217 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"image-registry-operator-tls\" (UniqueName: \"kubernetes.io/secret/2869c0eb-c1c6-4db7-9cf5-17e3300fa14d-image-registry-operator-tls\") pod \"cluster-image-registry-operator-dc59b4c8b-4fn9j\" (UID: \"2869c0eb-c1c6-4db7-9cf5-17e3300fa14d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4fn9j" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.627171 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"control-plane-machine-set-operator-tls\" (UniqueName: \"kubernetes.io/secret/2352fbc0-94ff-4b0c-b499-d16b94c279df-control-plane-machine-set-operator-tls\") pod \"control-plane-machine-set-operator-78cbb6b69f-bj72v\" (UID: \"2352fbc0-94ff-4b0c-b499-d16b94c279df\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bj72v" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.628662 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/5e5a84d1-542e-4b58-b991-611bd3202bcf-serving-cert\") pod \"etcd-operator-b45778765-pzgdk\" (UID: \"5e5a84d1-542e-4b58-b991-611bd3202bcf\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pzgdk" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.628854 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"default-certificate\" (UniqueName: \"kubernetes.io/secret/e6184ca8-dc6e-4d34-ad0e-87c7107acb82-default-certificate\") pod \"router-default-5444994796-7vhkr\" (UID: \"e6184ca8-dc6e-4d34-ad0e-87c7107acb82\") " pod="openshift-ingress/router-default-5444994796-7vhkr" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.630582 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/49fe2194-f940-4e0a-a214-fc6080068163-serving-cert\") pod \"service-ca-operator-777779d784-kjggj\" (UID: \"49fe2194-f940-4e0a-a214-fc6080068163\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-kjggj" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.630787 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2b9e45c7-edbb-4cbb-9ee2-ae7efb61fbb3-metrics-tls\") pod \"ingress-operator-5b745b69d9-z6v9h\" (UID: \"2b9e45c7-edbb-4cbb-9ee2-ae7efb61fbb3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-z6v9h" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.631693 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/b265c22f-ab92-48e4-9c3e-db6ea6956f6f-trusted-ca\") pod \"console-operator-58897d9998-z7w95\" (UID: \"b265c22f-ab92-48e4-9c3e-db6ea6956f6f\") " pod="openshift-console-operator/console-operator-58897d9998-z7w95" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.633507 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"stats-auth\" (UniqueName: \"kubernetes.io/secret/e6184ca8-dc6e-4d34-ad0e-87c7107acb82-stats-auth\") pod \"router-default-5444994796-7vhkr\" (UID: \"e6184ca8-dc6e-4d34-ad0e-87c7107acb82\") " pod="openshift-ingress/router-default-5444994796-7vhkr" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.635018 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etcd-client\" (UniqueName: \"kubernetes.io/secret/5e5a84d1-542e-4b58-b991-611bd3202bcf-etcd-client\") pod \"etcd-operator-b45778765-pzgdk\" (UID: \"5e5a84d1-542e-4b58-b991-611bd3202bcf\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pzgdk" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.635105 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver/apiserver-76f77b778f-6jfc4"] Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.635555 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/4dc52af7-648a-40e7-bdae-50672f5143dc-serving-cert\") pod \"kube-apiserver-operator-766d6c64bb-gc7vl\" (UID: \"4dc52af7-648a-40e7-bdae-50672f5143dc\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gc7vl" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.635744 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-certs\" (UniqueName: \"kubernetes.io/secret/09e21136-b8c7-4680-ab71-73a86766a7b8-webhook-certs\") pod \"multus-admission-controller-857f4d67dd-d24n2\" (UID: \"09e21136-b8c7-4680-ab71-73a86766a7b8\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-d24n2" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.635853 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/e6184ca8-dc6e-4d34-ad0e-87c7107acb82-metrics-certs\") pod \"router-default-5444994796-7vhkr\" (UID: \"e6184ca8-dc6e-4d34-ad0e-87c7107acb82\") " pod="openshift-ingress/router-default-5444994796-7vhkr" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.636132 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/db40ba09-a72f-43d5-8037-e71c96e2bf5b-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-zkbhv\" (UID: \"db40ba09-a72f-43d5-8037-e71c96e2bf5b\") " pod="openshift-marketplace/marketplace-operator-79b997595-zkbhv" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.636514 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/d08b0c90-3166-4919-88b3-345f5be3efe7-serving-cert\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6m4k7\" (UID: \"d08b0c90-3166-4919-88b3-345f5be3efe7\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6m4k7" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.641410 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2b9e45c7-edbb-4cbb-9ee2-ae7efb61fbb3-bound-sa-token\") pod \"ingress-operator-5b745b69d9-z6v9h\" (UID: \"2b9e45c7-edbb-4cbb-9ee2-ae7efb61fbb3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-z6v9h" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.660131 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qtpms\" (UniqueName: \"kubernetes.io/projected/a79948db-9342-439c-95ae-128c8bd0524d-kube-api-access-qtpms\") pod \"openshift-controller-manager-operator-756b6f6bc6-qc4wh\" (UID: \"a79948db-9342-439c-95ae-128c8bd0524d\") " pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qc4wh" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.664155 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vjnjc"] Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.666351 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-oauth-apiserver/apiserver-7bbb656c7d-fx7lk"] Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.678937 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ngczm\" (UniqueName: \"kubernetes.io/projected/70b489ae-f74b-4e9f-a5b2-71ec1c1f002c-kube-api-access-ngczm\") pod \"dns-operator-744455d44c-xs8b7\" (UID: \"70b489ae-f74b-4e9f-a5b2-71ec1c1f002c\") " pod="openshift-dns-operator/dns-operator-744455d44c-xs8b7" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.697528 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns-operator/dns-operator-744455d44c-xs8b7" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.701686 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbrg4\" (UniqueName: \"kubernetes.io/projected/2c8b7f32-4f80-4f95-b361-7d4687416711-kube-api-access-lbrg4\") pod \"image-pruner-29554560-slpk9\" (UID: \"2c8b7f32-4f80-4f95-b361-7d4687416711\") " pod="openshift-image-registry/image-pruner-29554560-slpk9" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.718137 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qc4wh" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.719695 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/2869c0eb-c1c6-4db7-9cf5-17e3300fa14d-bound-sa-token\") pod \"cluster-image-registry-operator-dc59b4c8b-4fn9j\" (UID: \"2869c0eb-c1c6-4db7-9cf5-17e3300fa14d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4fn9j" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.721801 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 00:10:17 crc kubenswrapper[4948]: E0312 00:10:17.722020 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 00:10:18.222002256 +0000 UTC m=+217.677605994 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.722121 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.722251 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xj5p2\" (UniqueName: \"kubernetes.io/projected/bd97cad6-0d6d-4bec-9d2f-e17b03b220f6-kube-api-access-xj5p2\") pod \"machine-config-controller-84d6567774-lkhdg\" (UID: \"bd97cad6-0d6d-4bec-9d2f-e17b03b220f6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lkhdg" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.722314 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sjmkx\" (UniqueName: \"kubernetes.io/projected/12c99e25-86bd-4923-ab44-55e571661c0f-kube-api-access-sjmkx\") pod \"olm-operator-6b444d44fb-h6z8v\" (UID: \"12c99e25-86bd-4923-ab44-55e571661c0f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h6z8v" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.722347 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/4ba12474-c71f-429c-bcdf-9d74eb0b4b7a-csi-data-dir\") pod \"csi-hostpathplugin-5q9rs\" (UID: \"4ba12474-c71f-429c-bcdf-9d74eb0b4b7a\") " pod="hostpath-provisioner/csi-hostpathplugin-5q9rs" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.722383 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/d33d5db8-961e-4c7a-950e-cb3f88a3f9b0-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-jlrln\" (UID: \"d33d5db8-961e-4c7a-950e-cb3f88a3f9b0\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jlrln" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.722420 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7d1dc3d3-060c-4aba-aa54-0a3ea767f299-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-v59sv\" (UID: \"7d1dc3d3-060c-4aba-aa54-0a3ea767f299\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-v59sv" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.722451 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/7599787e-ea15-40ca-b365-fdf406e3c337-srv-cert\") pod \"catalog-operator-68c6474976-dc6wb\" (UID: \"7599787e-ea15-40ca-b365-fdf406e3c337\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dc6wb" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.722478 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d4x4t\" (UniqueName: \"kubernetes.io/projected/2851d901-f861-432b-a61e-5e0391f1b2b0-kube-api-access-d4x4t\") pod \"dns-default-prqnt\" (UID: \"2851d901-f861-432b-a61e-5e0391f1b2b0\") " pod="openshift-dns/dns-default-prqnt" Mar 12 00:10:17 crc kubenswrapper[4948]: E0312 00:10:17.722497 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 00:10:18.222486577 +0000 UTC m=+217.678090315 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g4kcb" (UID: "af75ce0a-d6bb-48db-a422-08b1d87e1d3f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.722523 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5b801839-e39a-48a9-9eee-1bb95df5fbfb-config-volume\") pod \"collect-profiles-29554560-8mgv9\" (UID: \"5b801839-e39a-48a9-9eee-1bb95df5fbfb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29554560-8mgv9" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.722551 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/4ba12474-c71f-429c-bcdf-9d74eb0b4b7a-mountpoint-dir\") pod \"csi-hostpathplugin-5q9rs\" (UID: \"4ba12474-c71f-429c-bcdf-9d74eb0b4b7a\") " pod="hostpath-provisioner/csi-hostpathplugin-5q9rs" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.722610 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d1dc3d3-060c-4aba-aa54-0a3ea767f299-config\") pod \"kube-controller-manager-operator-78b949d7b-v59sv\" (UID: \"7d1dc3d3-060c-4aba-aa54-0a3ea767f299\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-v59sv" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.722634 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3e538a31-98b5-4353-8038-4a7580be89dd-webhook-cert\") pod \"packageserver-d55dfcdfc-5qjdn\" (UID: \"3e538a31-98b5-4353-8038-4a7580be89dd\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5qjdn" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.722658 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/cadf90b6-df8c-48fa-b878-68f2c7d7e49f-node-bootstrap-token\") pod \"machine-config-server-rvhq7\" (UID: \"cadf90b6-df8c-48fa-b878-68f2c7d7e49f\") " pod="openshift-machine-config-operator/machine-config-server-rvhq7" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.722714 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/bd97cad6-0d6d-4bec-9d2f-e17b03b220f6-proxy-tls\") pod \"machine-config-controller-84d6567774-lkhdg\" (UID: \"bd97cad6-0d6d-4bec-9d2f-e17b03b220f6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lkhdg" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.722755 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2851d901-f861-432b-a61e-5e0391f1b2b0-metrics-tls\") pod \"dns-default-prqnt\" (UID: \"2851d901-f861-432b-a61e-5e0391f1b2b0\") " pod="openshift-dns/dns-default-prqnt" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.722784 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/7599787e-ea15-40ca-b365-fdf406e3c337-profile-collector-cert\") pod \"catalog-operator-68c6474976-dc6wb\" (UID: \"7599787e-ea15-40ca-b365-fdf406e3c337\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dc6wb" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.722809 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/141bb112-1f07-4db5-9a4e-b7d94f12c5e2-signing-cabundle\") pod \"service-ca-9c57cc56f-q79m4\" (UID: \"141bb112-1f07-4db5-9a4e-b7d94f12c5e2\") " pod="openshift-service-ca/service-ca-9c57cc56f-q79m4" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.722837 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"csi-data-dir\" (UniqueName: \"kubernetes.io/host-path/4ba12474-c71f-429c-bcdf-9d74eb0b4b7a-csi-data-dir\") pod \"csi-hostpathplugin-5q9rs\" (UID: \"4ba12474-c71f-429c-bcdf-9d74eb0b4b7a\") " pod="hostpath-provisioner/csi-hostpathplugin-5q9rs" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.722844 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/4ba12474-c71f-429c-bcdf-9d74eb0b4b7a-socket-dir\") pod \"csi-hostpathplugin-5q9rs\" (UID: \"4ba12474-c71f-429c-bcdf-9d74eb0b4b7a\") " pod="hostpath-provisioner/csi-hostpathplugin-5q9rs" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.722868 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/12c99e25-86bd-4923-ab44-55e571661c0f-srv-cert\") pod \"olm-operator-6b444d44fb-h6z8v\" (UID: \"12c99e25-86bd-4923-ab44-55e571661c0f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h6z8v" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.722893 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"certs\" (UniqueName: \"kubernetes.io/secret/cadf90b6-df8c-48fa-b878-68f2c7d7e49f-certs\") pod \"machine-config-server-rvhq7\" (UID: \"cadf90b6-df8c-48fa-b878-68f2c7d7e49f\") " pod="openshift-machine-config-operator/machine-config-server-rvhq7" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.722916 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kn2xg\" (UniqueName: \"kubernetes.io/projected/d33d5db8-961e-4c7a-950e-cb3f88a3f9b0-kube-api-access-kn2xg\") pod \"package-server-manager-789f6589d5-jlrln\" (UID: \"d33d5db8-961e-4c7a-950e-cb3f88a3f9b0\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jlrln" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.722935 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert\" (UniqueName: \"kubernetes.io/secret/15828966-e769-43d4-b4a3-7a509fa72814-cert\") pod \"ingress-canary-mdhf9\" (UID: \"15828966-e769-43d4-b4a3-7a509fa72814\") " pod="openshift-ingress-canary/ingress-canary-mdhf9" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.722954 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/12c99e25-86bd-4923-ab44-55e571661c0f-profile-collector-cert\") pod \"olm-operator-6b444d44fb-h6z8v\" (UID: \"12c99e25-86bd-4923-ab44-55e571661c0f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h6z8v" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.722979 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5b801839-e39a-48a9-9eee-1bb95df5fbfb-secret-volume\") pod \"collect-profiles-29554560-8mgv9\" (UID: \"5b801839-e39a-48a9-9eee-1bb95df5fbfb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29554560-8mgv9" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.723000 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/4ba12474-c71f-429c-bcdf-9d74eb0b4b7a-plugins-dir\") pod \"csi-hostpathplugin-5q9rs\" (UID: \"4ba12474-c71f-429c-bcdf-9d74eb0b4b7a\") " pod="hostpath-provisioner/csi-hostpathplugin-5q9rs" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.723031 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6dvsl\" (UniqueName: \"kubernetes.io/projected/2fb324d6-44d4-440d-ae60-a5356a3a057f-kube-api-access-6dvsl\") pod \"auto-csr-approver-29554570-2b2hb\" (UID: \"2fb324d6-44d4-440d-ae60-a5356a3a057f\") " pod="openshift-infra/auto-csr-approver-29554570-2b2hb" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.723063 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-bm4bs\" (UniqueName: \"kubernetes.io/projected/7599787e-ea15-40ca-b365-fdf406e3c337-kube-api-access-bm4bs\") pod \"catalog-operator-68c6474976-dc6wb\" (UID: \"7599787e-ea15-40ca-b365-fdf406e3c337\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dc6wb" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.723091 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8btlp\" (UniqueName: \"kubernetes.io/projected/cadf90b6-df8c-48fa-b878-68f2c7d7e49f-kube-api-access-8btlp\") pod \"machine-config-server-rvhq7\" (UID: \"cadf90b6-df8c-48fa-b878-68f2c7d7e49f\") " pod="openshift-machine-config-operator/machine-config-server-rvhq7" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.723129 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2851d901-f861-432b-a61e-5e0391f1b2b0-config-volume\") pod \"dns-default-prqnt\" (UID: \"2851d901-f861-432b-a61e-5e0391f1b2b0\") " pod="openshift-dns/dns-default-prqnt" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.723157 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3e538a31-98b5-4353-8038-4a7580be89dd-apiservice-cert\") pod \"packageserver-d55dfcdfc-5qjdn\" (UID: \"3e538a31-98b5-4353-8038-4a7580be89dd\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5qjdn" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.723185 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/3e538a31-98b5-4353-8038-4a7580be89dd-tmpfs\") pod \"packageserver-d55dfcdfc-5qjdn\" (UID: \"3e538a31-98b5-4353-8038-4a7580be89dd\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5qjdn" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.723206 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fzbkp\" (UniqueName: \"kubernetes.io/projected/5b801839-e39a-48a9-9eee-1bb95df5fbfb-kube-api-access-fzbkp\") pod \"collect-profiles-29554560-8mgv9\" (UID: \"5b801839-e39a-48a9-9eee-1bb95df5fbfb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29554560-8mgv9" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.723230 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-9vs6g\" (UniqueName: \"kubernetes.io/projected/141bb112-1f07-4db5-9a4e-b7d94f12c5e2-kube-api-access-9vs6g\") pod \"service-ca-9c57cc56f-q79m4\" (UID: \"141bb112-1f07-4db5-9a4e-b7d94f12c5e2\") " pod="openshift-service-ca/service-ca-9c57cc56f-q79m4" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.723267 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7d1dc3d3-060c-4aba-aa54-0a3ea767f299-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-v59sv\" (UID: \"7d1dc3d3-060c-4aba-aa54-0a3ea767f299\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-v59sv" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.723290 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-q4687\" (UniqueName: \"kubernetes.io/projected/3e538a31-98b5-4353-8038-4a7580be89dd-kube-api-access-q4687\") pod \"packageserver-d55dfcdfc-5qjdn\" (UID: \"3e538a31-98b5-4353-8038-4a7580be89dd\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5qjdn" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.723362 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/bd97cad6-0d6d-4bec-9d2f-e17b03b220f6-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-lkhdg\" (UID: \"bd97cad6-0d6d-4bec-9d2f-e17b03b220f6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lkhdg" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.723386 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/4ba12474-c71f-429c-bcdf-9d74eb0b4b7a-registration-dir\") pod \"csi-hostpathplugin-5q9rs\" (UID: \"4ba12474-c71f-429c-bcdf-9d74eb0b4b7a\") " pod="hostpath-provisioner/csi-hostpathplugin-5q9rs" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.723409 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rs2ns\" (UniqueName: \"kubernetes.io/projected/4ba12474-c71f-429c-bcdf-9d74eb0b4b7a-kube-api-access-rs2ns\") pod \"csi-hostpathplugin-5q9rs\" (UID: \"4ba12474-c71f-429c-bcdf-9d74eb0b4b7a\") " pod="hostpath-provisioner/csi-hostpathplugin-5q9rs" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.723429 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/141bb112-1f07-4db5-9a4e-b7d94f12c5e2-signing-key\") pod \"service-ca-9c57cc56f-q79m4\" (UID: \"141bb112-1f07-4db5-9a4e-b7d94f12c5e2\") " pod="openshift-service-ca/service-ca-9c57cc56f-q79m4" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.723451 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-hsk7b\" (UniqueName: \"kubernetes.io/projected/15828966-e769-43d4-b4a3-7a509fa72814-kube-api-access-hsk7b\") pod \"ingress-canary-mdhf9\" (UID: \"15828966-e769-43d4-b4a3-7a509fa72814\") " pod="openshift-ingress-canary/ingress-canary-mdhf9" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.724798 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5b801839-e39a-48a9-9eee-1bb95df5fbfb-config-volume\") pod \"collect-profiles-29554560-8mgv9\" (UID: \"5b801839-e39a-48a9-9eee-1bb95df5fbfb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29554560-8mgv9" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.725920 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mountpoint-dir\" (UniqueName: \"kubernetes.io/host-path/4ba12474-c71f-429c-bcdf-9d74eb0b4b7a-mountpoint-dir\") pod \"csi-hostpathplugin-5q9rs\" (UID: \"4ba12474-c71f-429c-bcdf-9d74eb0b4b7a\") " pod="hostpath-provisioner/csi-hostpathplugin-5q9rs" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.726616 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/2851d901-f861-432b-a61e-5e0391f1b2b0-config-volume\") pod \"dns-default-prqnt\" (UID: \"2851d901-f861-432b-a61e-5e0391f1b2b0\") " pod="openshift-dns/dns-default-prqnt" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.727427 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"package-server-manager-serving-cert\" (UniqueName: \"kubernetes.io/secret/d33d5db8-961e-4c7a-950e-cb3f88a3f9b0-package-server-manager-serving-cert\") pod \"package-server-manager-789f6589d5-jlrln\" (UID: \"d33d5db8-961e-4c7a-950e-cb3f88a3f9b0\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jlrln" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.728223 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/7599787e-ea15-40ca-b365-fdf406e3c337-srv-cert\") pod \"catalog-operator-68c6474976-dc6wb\" (UID: \"7599787e-ea15-40ca-b365-fdf406e3c337\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dc6wb" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.728225 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-cabundle\" (UniqueName: \"kubernetes.io/configmap/141bb112-1f07-4db5-9a4e-b7d94f12c5e2-signing-cabundle\") pod \"service-ca-9c57cc56f-q79m4\" (UID: \"141bb112-1f07-4db5-9a4e-b7d94f12c5e2\") " pod="openshift-service-ca/service-ca-9c57cc56f-q79m4" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.729225 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/12c99e25-86bd-4923-ab44-55e571661c0f-profile-collector-cert\") pod \"olm-operator-6b444d44fb-h6z8v\" (UID: \"12c99e25-86bd-4923-ab44-55e571661c0f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h6z8v" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.729654 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"socket-dir\" (UniqueName: \"kubernetes.io/host-path/4ba12474-c71f-429c-bcdf-9d74eb0b4b7a-socket-dir\") pod \"csi-hostpathplugin-5q9rs\" (UID: \"4ba12474-c71f-429c-bcdf-9d74eb0b4b7a\") " pod="hostpath-provisioner/csi-hostpathplugin-5q9rs" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.729675 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/7d1dc3d3-060c-4aba-aa54-0a3ea767f299-config\") pod \"kube-controller-manager-operator-78b949d7b-v59sv\" (UID: \"7d1dc3d3-060c-4aba-aa54-0a3ea767f299\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-v59sv" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.729875 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/downloads-7954f5f757-xdh6w" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.730243 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registration-dir\" (UniqueName: \"kubernetes.io/host-path/4ba12474-c71f-429c-bcdf-9d74eb0b4b7a-registration-dir\") pod \"csi-hostpathplugin-5q9rs\" (UID: \"4ba12474-c71f-429c-bcdf-9d74eb0b4b7a\") " pod="hostpath-provisioner/csi-hostpathplugin-5q9rs" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.730335 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-tls\" (UniqueName: \"kubernetes.io/secret/2851d901-f861-432b-a61e-5e0391f1b2b0-metrics-tls\") pod \"dns-default-prqnt\" (UID: \"2851d901-f861-432b-a61e-5e0391f1b2b0\") " pod="openshift-dns/dns-default-prqnt" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.730341 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"plugins-dir\" (UniqueName: \"kubernetes.io/host-path/4ba12474-c71f-429c-bcdf-9d74eb0b4b7a-plugins-dir\") pod \"csi-hostpathplugin-5q9rs\" (UID: \"4ba12474-c71f-429c-bcdf-9d74eb0b4b7a\") " pod="hostpath-provisioner/csi-hostpathplugin-5q9rs" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.730642 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-bootstrap-token\" (UniqueName: \"kubernetes.io/secret/cadf90b6-df8c-48fa-b878-68f2c7d7e49f-node-bootstrap-token\") pod \"machine-config-server-rvhq7\" (UID: \"cadf90b6-df8c-48fa-b878-68f2c7d7e49f\") " pod="openshift-machine-config-operator/machine-config-server-rvhq7" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.730892 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/7d1dc3d3-060c-4aba-aa54-0a3ea767f299-serving-cert\") pod \"kube-controller-manager-operator-78b949d7b-v59sv\" (UID: \"7d1dc3d3-060c-4aba-aa54-0a3ea767f299\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-v59sv" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.731047 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"mcc-auth-proxy-config\" (UniqueName: \"kubernetes.io/configmap/bd97cad6-0d6d-4bec-9d2f-e17b03b220f6-mcc-auth-proxy-config\") pod \"machine-config-controller-84d6567774-lkhdg\" (UID: \"bd97cad6-0d6d-4bec-9d2f-e17b03b220f6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lkhdg" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.731138 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmpfs\" (UniqueName: \"kubernetes.io/empty-dir/3e538a31-98b5-4353-8038-4a7580be89dd-tmpfs\") pod \"packageserver-d55dfcdfc-5qjdn\" (UID: \"3e538a31-98b5-4353-8038-4a7580be89dd\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5qjdn" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.731485 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"signing-key\" (UniqueName: \"kubernetes.io/secret/141bb112-1f07-4db5-9a4e-b7d94f12c5e2-signing-key\") pod \"service-ca-9c57cc56f-q79m4\" (UID: \"141bb112-1f07-4db5-9a4e-b7d94f12c5e2\") " pod="openshift-service-ca/service-ca-9c57cc56f-q79m4" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.731655 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-tls\" (UniqueName: \"kubernetes.io/secret/bd97cad6-0d6d-4bec-9d2f-e17b03b220f6-proxy-tls\") pod \"machine-config-controller-84d6567774-lkhdg\" (UID: \"bd97cad6-0d6d-4bec-9d2f-e17b03b220f6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lkhdg" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.732555 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/3e538a31-98b5-4353-8038-4a7580be89dd-webhook-cert\") pod \"packageserver-d55dfcdfc-5qjdn\" (UID: \"3e538a31-98b5-4353-8038-4a7580be89dd\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5qjdn" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.740222 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gfgnq\" (UniqueName: \"kubernetes.io/projected/892017c6-01a1-4798-9d23-fff4f44c6932-kube-api-access-gfgnq\") pod \"console-f9d7485db-qq8gg\" (UID: \"892017c6-01a1-4798-9d23-fff4f44c6932\") " pod="openshift-console/console-f9d7485db-qq8gg" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.741131 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"certs\" (UniqueName: \"kubernetes.io/secret/cadf90b6-df8c-48fa-b878-68f2c7d7e49f-certs\") pod \"machine-config-server-rvhq7\" (UID: \"cadf90b6-df8c-48fa-b878-68f2c7d7e49f\") " pod="openshift-machine-config-operator/machine-config-server-rvhq7" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.744875 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/3e538a31-98b5-4353-8038-4a7580be89dd-apiservice-cert\") pod \"packageserver-d55dfcdfc-5qjdn\" (UID: \"3e538a31-98b5-4353-8038-4a7580be89dd\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5qjdn" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.746428 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29554560-slpk9" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.746433 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert\" (UniqueName: \"kubernetes.io/secret/15828966-e769-43d4-b4a3-7a509fa72814-cert\") pod \"ingress-canary-mdhf9\" (UID: \"15828966-e769-43d4-b4a3-7a509fa72814\") " pod="openshift-ingress-canary/ingress-canary-mdhf9" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.746793 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"srv-cert\" (UniqueName: \"kubernetes.io/secret/12c99e25-86bd-4923-ab44-55e571661c0f-srv-cert\") pod \"olm-operator-6b444d44fb-h6z8v\" (UID: \"12c99e25-86bd-4923-ab44-55e571661c0f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h6z8v" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.747688 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5b801839-e39a-48a9-9eee-1bb95df5fbfb-secret-volume\") pod \"collect-profiles-29554560-8mgv9\" (UID: \"5b801839-e39a-48a9-9eee-1bb95df5fbfb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29554560-8mgv9" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.753114 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"profile-collector-cert\" (UniqueName: \"kubernetes.io/secret/7599787e-ea15-40ca-b365-fdf406e3c337-profile-collector-cert\") pod \"catalog-operator-68c6474976-dc6wb\" (UID: \"7599787e-ea15-40ca-b365-fdf406e3c337\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dc6wb" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.759963 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/d08b0c90-3166-4919-88b3-345f5be3efe7-kube-api-access\") pod \"openshift-kube-scheduler-operator-5fdd9b5758-6m4k7\" (UID: \"d08b0c90-3166-4919-88b3-345f5be3efe7\") " pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6m4k7" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.760339 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console/console-f9d7485db-qq8gg" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.766461 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6m4k7" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.783203 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7kscx\" (UniqueName: \"kubernetes.io/projected/49fe2194-f940-4e0a-a214-fc6080068163-kube-api-access-7kscx\") pod \"service-ca-operator-777779d784-kjggj\" (UID: \"49fe2194-f940-4e0a-a214-fc6080068163\") " pod="openshift-service-ca-operator/service-ca-operator-777779d784-kjggj" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.801038 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7k6hs\" (UniqueName: \"kubernetes.io/projected/09e21136-b8c7-4680-ab71-73a86766a7b8-kube-api-access-7k6hs\") pod \"multus-admission-controller-857f4d67dd-d24n2\" (UID: \"09e21136-b8c7-4680-ab71-73a86766a7b8\") " pod="openshift-multus/multus-admission-controller-857f4d67dd-d24n2" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.813562 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/multus-admission-controller-857f4d67dd-d24n2" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.830745 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 00:10:17 crc kubenswrapper[4948]: E0312 00:10:17.831074 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 00:10:18.331045526 +0000 UTC m=+217.786649254 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.831462 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:17 crc kubenswrapper[4948]: E0312 00:10:17.831883 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 00:10:18.331876194 +0000 UTC m=+217.787479932 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g4kcb" (UID: "af75ce0a-d6bb-48db-a422-08b1d87e1d3f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.832016 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4q42m\" (UniqueName: \"kubernetes.io/projected/b265c22f-ab92-48e4-9c3e-db6ea6956f6f-kube-api-access-4q42m\") pod \"console-operator-58897d9998-z7w95\" (UID: \"b265c22f-ab92-48e4-9c3e-db6ea6956f6f\") " pod="openshift-console-operator/console-operator-58897d9998-z7w95" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.842050 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca-operator/service-ca-operator-777779d784-kjggj" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.844967 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-r5cbw\" (UniqueName: \"kubernetes.io/projected/677936de-2686-46cf-99e9-cae2355c42b4-kube-api-access-r5cbw\") pod \"migrator-59844c95c7-srr9m\" (UID: \"677936de-2686-46cf-99e9-cae2355c42b4\") " pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-srr9m" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.865131 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rjspx\" (UniqueName: \"kubernetes.io/projected/1196fdb1-9c09-4982-9c37-35f46fabc1c5-kube-api-access-rjspx\") pod \"machine-config-operator-74547568cd-9k744\" (UID: \"1196fdb1-9c09-4982-9c37-35f46fabc1c5\") " pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9k744" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.884809 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wkr7n\" (UniqueName: \"kubernetes.io/projected/1666a214-546a-4063-a627-e990781797c8-kube-api-access-wkr7n\") pod \"kube-storage-version-migrator-operator-b67b599dd-kvh9w\" (UID: \"1666a214-546a-4063-a627-e990781797c8\") " pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kvh9w" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.901419 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-k69zs\" (UniqueName: \"kubernetes.io/projected/2869c0eb-c1c6-4db7-9cf5-17e3300fa14d-kube-api-access-k69zs\") pod \"cluster-image-registry-operator-dc59b4c8b-4fn9j\" (UID: \"2869c0eb-c1c6-4db7-9cf5-17e3300fa14d\") " pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4fn9j" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.922550 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-htmnx\" (UniqueName: \"kubernetes.io/projected/5e5a84d1-542e-4b58-b991-611bd3202bcf-kube-api-access-htmnx\") pod \"etcd-operator-b45778765-pzgdk\" (UID: \"5e5a84d1-542e-4b58-b991-611bd3202bcf\") " pod="openshift-etcd-operator/etcd-operator-b45778765-pzgdk" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.932698 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 00:10:17 crc kubenswrapper[4948]: E0312 00:10:17.933115 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 00:10:18.4330911 +0000 UTC m=+217.888694838 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.933181 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:17 crc kubenswrapper[4948]: E0312 00:10:17.934252 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 00:10:18.434238707 +0000 UTC m=+217.889842445 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g4kcb" (UID: "af75ce0a-d6bb-48db-a422-08b1d87e1d3f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.946384 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5t7dj\" (UniqueName: \"kubernetes.io/projected/e6184ca8-dc6e-4d34-ad0e-87c7107acb82-kube-api-access-5t7dj\") pod \"router-default-5444994796-7vhkr\" (UID: \"e6184ca8-dc6e-4d34-ad0e-87c7107acb82\") " pod="openshift-ingress/router-default-5444994796-7vhkr" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.964125 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jl7d7\" (UniqueName: \"kubernetes.io/projected/af75ce0a-d6bb-48db-a422-08b1d87e1d3f-kube-api-access-jl7d7\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.984106 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qm88d\" (UniqueName: \"kubernetes.io/projected/2352fbc0-94ff-4b0c-b499-d16b94c279df-kube-api-access-qm88d\") pod \"control-plane-machine-set-operator-78cbb6b69f-bj72v\" (UID: \"2352fbc0-94ff-4b0c-b499-d16b94c279df\") " pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bj72v" Mar 12 00:10:17 crc kubenswrapper[4948]: I0312 00:10:17.997008 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns-operator/dns-operator-744455d44c-xs8b7"] Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.006629 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/af75ce0a-d6bb-48db-a422-08b1d87e1d3f-bound-sa-token\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.011617 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4fn9j" Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.021954 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g7gzh\" (UniqueName: \"kubernetes.io/projected/db40ba09-a72f-43d5-8037-e71c96e2bf5b-kube-api-access-g7gzh\") pod \"marketplace-operator-79b997595-zkbhv\" (UID: \"db40ba09-a72f-43d5-8037-e71c96e2bf5b\") " pod="openshift-marketplace/marketplace-operator-79b997595-zkbhv" Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.028355 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/downloads-7954f5f757-xdh6w"] Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.032780 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-etcd-operator/etcd-operator-b45778765-pzgdk" Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.034094 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 00:10:18 crc kubenswrapper[4948]: E0312 00:10:18.034484 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 00:10:18.534469521 +0000 UTC m=+217.990073259 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.040540 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-console-operator/console-operator-58897d9998-z7w95" Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.042609 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/4dc52af7-648a-40e7-bdae-50672f5143dc-kube-api-access\") pod \"kube-apiserver-operator-766d6c64bb-gc7vl\" (UID: \"4dc52af7-648a-40e7-bdae-50672f5143dc\") " pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gc7vl" Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.063971 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fj9zg\" (UniqueName: \"kubernetes.io/projected/2b9e45c7-edbb-4cbb-9ee2-ae7efb61fbb3-kube-api-access-fj9zg\") pod \"ingress-operator-5b745b69d9-z6v9h\" (UID: \"2b9e45c7-edbb-4cbb-9ee2-ae7efb61fbb3\") " pod="openshift-ingress-operator/ingress-operator-5b745b69d9-z6v9h" Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.072875 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-z6v9h" Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.079182 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-pruner-29554560-slpk9"] Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.079566 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kvh9w" Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.084668 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-srr9m" Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.085021 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qk6mp\" (UniqueName: \"kubernetes.io/projected/29cf53b1-0215-4dd4-a19f-24a48d4ded9d-kube-api-access-qk6mp\") pod \"cluster-samples-operator-665b6dd947-rfxxc\" (UID: \"29cf53b1-0215-4dd4-a19f-24a48d4ded9d\") " pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rfxxc" Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.090867 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress/router-default-5444994796-7vhkr" Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.099710 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bj72v" Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.105374 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gc7vl" Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.106003 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qc4wh"] Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.119663 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9k744" Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.125600 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-zkbhv" Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.129604 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sjmkx\" (UniqueName: \"kubernetes.io/projected/12c99e25-86bd-4923-ab44-55e571661c0f-kube-api-access-sjmkx\") pod \"olm-operator-6b444d44fb-h6z8v\" (UID: \"12c99e25-86bd-4923-ab44-55e571661c0f\") " pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h6z8v" Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.135379 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:18 crc kubenswrapper[4948]: E0312 00:10:18.135952 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 00:10:18.635939993 +0000 UTC m=+218.091543731 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g4kcb" (UID: "af75ce0a-d6bb-48db-a422-08b1d87e1d3f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.147489 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h6z8v" Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.151615 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xj5p2\" (UniqueName: \"kubernetes.io/projected/bd97cad6-0d6d-4bec-9d2f-e17b03b220f6-kube-api-access-xj5p2\") pod \"machine-config-controller-84d6567774-lkhdg\" (UID: \"bd97cad6-0d6d-4bec-9d2f-e17b03b220f6\") " pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lkhdg" Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.160974 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d4x4t\" (UniqueName: \"kubernetes.io/projected/2851d901-f861-432b-a61e-5e0391f1b2b0-kube-api-access-d4x4t\") pod \"dns-default-prqnt\" (UID: \"2851d901-f861-432b-a61e-5e0391f1b2b0\") " pod="openshift-dns/dns-default-prqnt" Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.187068 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-hsk7b\" (UniqueName: \"kubernetes.io/projected/15828966-e769-43d4-b4a3-7a509fa72814-kube-api-access-hsk7b\") pod \"ingress-canary-mdhf9\" (UID: \"15828966-e769-43d4-b4a3-7a509fa72814\") " pod="openshift-ingress-canary/ingress-canary-mdhf9" Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.201352 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6dvsl\" (UniqueName: \"kubernetes.io/projected/2fb324d6-44d4-440d-ae60-a5356a3a057f-kube-api-access-6dvsl\") pod \"auto-csr-approver-29554570-2b2hb\" (UID: \"2fb324d6-44d4-440d-ae60-a5356a3a057f\") " pod="openshift-infra/auto-csr-approver-29554570-2b2hb" Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.216947 4948 generic.go:334] "Generic (PLEG): container finished" podID="2040ed8d-ae32-4909-9161-bc7ffbe49edd" containerID="8ac72abffe605e3c26e37b76f64f492fd1fdf938c6cb1ed7260ffc1922d23cfd" exitCode=0 Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.217014 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-6jfc4" event={"ID":"2040ed8d-ae32-4909-9161-bc7ffbe49edd","Type":"ContainerDied","Data":"8ac72abffe605e3c26e37b76f64f492fd1fdf938c6cb1ed7260ffc1922d23cfd"} Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.217075 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-6jfc4" event={"ID":"2040ed8d-ae32-4909-9161-bc7ffbe49edd","Type":"ContainerStarted","Data":"42aca8266815fe638c8e6a18824edb72926bcedecc1ee257c31fea79eb12cf98"} Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.221929 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lkhdg" Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.231185 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kn2xg\" (UniqueName: \"kubernetes.io/projected/d33d5db8-961e-4c7a-950e-cb3f88a3f9b0-kube-api-access-kn2xg\") pod \"package-server-manager-789f6589d5-jlrln\" (UID: \"d33d5db8-961e-4c7a-950e-cb3f88a3f9b0\") " pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jlrln" Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.231901 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ingress-canary/ingress-canary-mdhf9" Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.236781 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 00:10:18 crc kubenswrapper[4948]: E0312 00:10:18.237176 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 00:10:18.73715976 +0000 UTC m=+218.192763498 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.238291 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-dns/dns-default-prqnt" Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.243741 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-9vs6g\" (UniqueName: \"kubernetes.io/projected/141bb112-1f07-4db5-9a4e-b7d94f12c5e2-kube-api-access-9vs6g\") pod \"service-ca-9c57cc56f-q79m4\" (UID: \"141bb112-1f07-4db5-9a4e-b7d94f12c5e2\") " pod="openshift-service-ca/service-ca-9c57cc56f-q79m4" Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.260788 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-bm4bs\" (UniqueName: \"kubernetes.io/projected/7599787e-ea15-40ca-b365-fdf406e3c337-kube-api-access-bm4bs\") pod \"catalog-operator-68c6474976-dc6wb\" (UID: \"7599787e-ea15-40ca-b365-fdf406e3c337\") " pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dc6wb" Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.263747 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vjnjc" event={"ID":"b36cdace-7dbb-4e8a-b0f0-6e93615612a8","Type":"ContainerStarted","Data":"466b58ade820ed0e650ce00d51de8ec2216cfa624ada2117014dc08f8c83660e"} Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.263797 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vjnjc" event={"ID":"b36cdace-7dbb-4e8a-b0f0-6e93615612a8","Type":"ContainerStarted","Data":"fc560786197f8f2d18c257457ef857cd4635e6278f19307cc40ce835cbe11079"} Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.270379 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-xs8b7" event={"ID":"70b489ae-f74b-4e9f-a5b2-71ec1c1f002c","Type":"ContainerStarted","Data":"4f75d5bb9a75d59641b59e852639dd5b8e54106a4984a07d1c358d11c96fdb31"} Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.278515 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-n4545" event={"ID":"a6da8984-cc2d-4b58-b6da-d7fc2d1e1725","Type":"ContainerStarted","Data":"f83fe96984e509a08cc9a04f8bbcc4cb5a1ff04197d1e753eb5bc11fde30fb92"} Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.278552 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-n4545" event={"ID":"a6da8984-cc2d-4b58-b6da-d7fc2d1e1725","Type":"ContainerStarted","Data":"6180b9ac487cbd1a227e1d2e55c90b5e78613c37a792d13440fe9f7635826981"} Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.281592 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qc4wh" event={"ID":"a79948db-9342-439c-95ae-128c8bd0524d","Type":"ContainerStarted","Data":"2f901f32d478f824b81c0b96cbbacbf56a728ab3085510fe2d2e0843150ba273"} Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.284514 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8btlp\" (UniqueName: \"kubernetes.io/projected/cadf90b6-df8c-48fa-b878-68f2c7d7e49f-kube-api-access-8btlp\") pod \"machine-config-server-rvhq7\" (UID: \"cadf90b6-df8c-48fa-b878-68f2c7d7e49f\") " pod="openshift-machine-config-operator/machine-config-server-rvhq7" Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.291025 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rp8sv" event={"ID":"60024b0b-f817-477f-83f1-e53cbc1e05b0","Type":"ContainerStarted","Data":"567bd37219624454e44e80dc908b61d93b5a4feeeecc47a009a525503103bc90"} Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.291074 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rp8sv" event={"ID":"60024b0b-f817-477f-83f1-e53cbc1e05b0","Type":"ContainerStarted","Data":"4795f512d0d3efb3414bf29827a15719ed2468da83504acccf6d4089a5361df6"} Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.291098 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rp8sv" Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.291593 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-etcd-operator/etcd-operator-b45778765-pzgdk"] Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.293547 4948 patch_prober.go:28] interesting pod/route-controller-manager-6576b87f9c-rp8sv container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" start-of-body= Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.293645 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rp8sv" podUID="60024b0b-f817-477f-83f1-e53cbc1e05b0" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.9:8443/healthz\": dial tcp 10.217.0.9:8443: connect: connection refused" Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.300590 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-9tpt4" event={"ID":"9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5","Type":"ContainerStarted","Data":"6d8adce161c66b17114e112a0fc6f4519e1429afd2849dbda606aa4048b581b2"} Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.300622 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-9tpt4" event={"ID":"9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5","Type":"ContainerStarted","Data":"500e98a6770d1487c1d9daeeb538a9a1984dab56b1e205303c446292ff336d69"} Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.300985 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-879f6c89f-9tpt4" Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.303086 4948 patch_prober.go:28] interesting pod/controller-manager-879f6c89f-9tpt4 container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" start-of-body= Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.303122 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-879f6c89f-9tpt4" podUID="9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.6:8443/healthz\": dial tcp 10.217.0.6:8443: connect: connection refused" Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.304175 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-5cwjt" event={"ID":"5c9bc3de-a78c-44d5-90ca-f57a625e125a","Type":"ContainerStarted","Data":"a757acedd97cf5627b89763fb7f1387ec5fe2c57cd5af9ee794c83a8f946bd15"} Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.304206 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication-operator/authentication-operator-69f744f599-5cwjt" event={"ID":"5c9bc3de-a78c-44d5-90ca-f57a625e125a","Type":"ContainerStarted","Data":"5465881d311a1cd2cadcc76214b50fa98f99af28ad0643385fc965b5ce1330bd"} Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.311934 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-q4687\" (UniqueName: \"kubernetes.io/projected/3e538a31-98b5-4353-8038-4a7580be89dd-kube-api-access-q4687\") pod \"packageserver-d55dfcdfc-5qjdn\" (UID: \"3e538a31-98b5-4353-8038-4a7580be89dd\") " pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5qjdn" Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.315027 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console/console-f9d7485db-qq8gg"] Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.324641 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fzbkp\" (UniqueName: \"kubernetes.io/projected/5b801839-e39a-48a9-9eee-1bb95df5fbfb-kube-api-access-fzbkp\") pod \"collect-profiles-29554560-8mgv9\" (UID: \"5b801839-e39a-48a9-9eee-1bb95df5fbfb\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29554560-8mgv9" Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.342650 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:18 crc kubenswrapper[4948]: E0312 00:10:18.344516 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 00:10:18.844503848 +0000 UTC m=+218.300107586 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g4kcb" (UID: "af75ce0a-d6bb-48db-a422-08b1d87e1d3f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.344784 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rs2ns\" (UniqueName: \"kubernetes.io/projected/4ba12474-c71f-429c-bcdf-9d74eb0b4b7a-kube-api-access-rs2ns\") pod \"csi-hostpathplugin-5q9rs\" (UID: \"4ba12474-c71f-429c-bcdf-9d74eb0b4b7a\") " pod="hostpath-provisioner/csi-hostpathplugin-5q9rs" Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.346895 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-xdh6w" event={"ID":"74ab81d8-55e8-47da-adb5-1f616a100a31","Type":"ContainerStarted","Data":"cf0729e2c1253c597af95aa9c363583375f85830d2b318c4cca1aef8b1affefe"} Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.353510 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rfxxc" Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.357397 4948 generic.go:334] "Generic (PLEG): container finished" podID="a12c201c-f0ce-4519-8873-e7ec122e7fec" containerID="fbefdcdfdc46bd7c212419a2e948f1ae4e4e3fbfed264efa61f8ed3ade09c3bb" exitCode=0 Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.357451 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-sg87j" event={"ID":"a12c201c-f0ce-4519-8873-e7ec122e7fec","Type":"ContainerDied","Data":"fbefdcdfdc46bd7c212419a2e948f1ae4e4e3fbfed264efa61f8ed3ade09c3bb"} Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.357475 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-sg87j" event={"ID":"a12c201c-f0ce-4519-8873-e7ec122e7fec","Type":"ContainerStarted","Data":"90552a3f2f2bb7f56ca4ecf50a05eee80969d30f8f65d65fd6ff55edefc20cd8"} Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.360104 4948 generic.go:334] "Generic (PLEG): container finished" podID="f980811d-53fd-4c64-96c5-19283afcbb1f" containerID="fe18bf07ee6b96dfa195a66ebff8f1659f38f45892944b874953cc4eb9577e7e" exitCode=0 Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.360155 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fx7lk" event={"ID":"f980811d-53fd-4c64-96c5-19283afcbb1f","Type":"ContainerDied","Data":"fe18bf07ee6b96dfa195a66ebff8f1659f38f45892944b874953cc4eb9577e7e"} Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.360178 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fx7lk" event={"ID":"f980811d-53fd-4c64-96c5-19283afcbb1f","Type":"ContainerStarted","Data":"937432f71dded835016574ba0236011fd8af157914e4435ee495fe496dffdf0f"} Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.370675 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7d1dc3d3-060c-4aba-aa54-0a3ea767f299-kube-api-access\") pod \"kube-controller-manager-operator-78b949d7b-v59sv\" (UID: \"7d1dc3d3-060c-4aba-aa54-0a3ea767f299\") " pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-v59sv" Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.395273 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-qsf8x" event={"ID":"f90934d6-9305-4513-8d84-cd6a6fcc4a59","Type":"ContainerStarted","Data":"c684b0e15925fe0fa95fd90d3a1677094513bf25ca14305b5365b19e206e8576"} Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.399544 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6m4k7"] Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.399576 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-console-operator/console-operator-58897d9998-z7w95"] Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.400649 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/multus-admission-controller-857f4d67dd-d24n2"] Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.401558 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-g7g94" event={"ID":"a544acdb-b758-4ea8-a43d-8270fc454901","Type":"ContainerStarted","Data":"f075c23d35c11e1ee8a48a099441654ea572e2d9ed93b1ca97990c7811a24aec"} Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.401584 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-g7g94" event={"ID":"a544acdb-b758-4ea8-a43d-8270fc454901","Type":"ContainerStarted","Data":"f21a87d35fc664563fafae537137b868e97bc6a227e9b814942f397207b5f35b"} Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.401593 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/machine-api-operator-5694c8668f-g7g94" event={"ID":"a544acdb-b758-4ea8-a43d-8270fc454901","Type":"ContainerStarted","Data":"bcaf812660f16c8387771b7d5ce1db8a404b09f01e854e8133878577c66e81f1"} Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.402753 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca-operator/service-ca-operator-777779d784-kjggj"] Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.443597 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 00:10:18 crc kubenswrapper[4948]: E0312 00:10:18.444597 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 00:10:18.944579528 +0000 UTC m=+218.400183266 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.455738 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29554560-8mgv9" Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.461552 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jlrln" Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.486876 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29554570-2b2hb" Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.498740 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5qjdn" Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.504708 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-v59sv" Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.511901 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dc6wb" Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.537356 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-service-ca/service-ca-9c57cc56f-q79m4" Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.545575 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="hostpath-provisioner/csi-hostpathplugin-5q9rs" Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.546827 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:18 crc kubenswrapper[4948]: E0312 00:10:18.547222 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 00:10:19.047207077 +0000 UTC m=+218.502810815 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g4kcb" (UID: "af75ce0a-d6bb-48db-a422-08b1d87e1d3f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.567777 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-operator/ingress-operator-5b745b69d9-z6v9h"] Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.577697 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-machine-config-operator/machine-config-server-rvhq7" Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.647227 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 00:10:18 crc kubenswrapper[4948]: E0312 00:10:18.647599 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 00:10:19.147585264 +0000 UTC m=+218.603189002 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.749400 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:18 crc kubenswrapper[4948]: E0312 00:10:18.749716 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 00:10:19.249701851 +0000 UTC m=+218.705305589 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g4kcb" (UID: "af75ce0a-d6bb-48db-a422-08b1d87e1d3f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.852702 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 00:10:18 crc kubenswrapper[4948]: E0312 00:10:18.853443 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 00:10:19.353428136 +0000 UTC m=+218.809031874 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.918123 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver-operator/openshift-apiserver-operator-796bbdcf4f-vjnjc" podStartSLOduration=179.918108558 podStartE2EDuration="2m59.918108558s" podCreationTimestamp="2026-03-12 00:07:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:10:18.918085318 +0000 UTC m=+218.373689056" watchObservedRunningTime="2026-03-12 00:10:18.918108558 +0000 UTC m=+218.373712296" Mar 12 00:10:18 crc kubenswrapper[4948]: I0312 00:10:18.954853 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:18 crc kubenswrapper[4948]: E0312 00:10:18.955373 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 00:10:19.455361139 +0000 UTC m=+218.910964877 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g4kcb" (UID: "af75ce0a-d6bb-48db-a422-08b1d87e1d3f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:19 crc kubenswrapper[4948]: I0312 00:10:19.055988 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 00:10:19 crc kubenswrapper[4948]: E0312 00:10:19.056897 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 00:10:19.556871762 +0000 UTC m=+219.012475500 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:19 crc kubenswrapper[4948]: I0312 00:10:19.065818 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:19 crc kubenswrapper[4948]: E0312 00:10:19.066318 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 00:10:19.566291993 +0000 UTC m=+219.021895731 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g4kcb" (UID: "af75ce0a-d6bb-48db-a422-08b1d87e1d3f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:19 crc kubenswrapper[4948]: I0312 00:10:19.086564 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bj72v"] Mar 12 00:10:19 crc kubenswrapper[4948]: I0312 00:10:19.133238 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-zkbhv"] Mar 12 00:10:19 crc kubenswrapper[4948]: I0312 00:10:19.156722 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gc7vl"] Mar 12 00:10:19 crc kubenswrapper[4948]: I0312 00:10:19.167606 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 00:10:19 crc kubenswrapper[4948]: E0312 00:10:19.167908 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 00:10:19.667894937 +0000 UTC m=+219.123498675 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:19 crc kubenswrapper[4948]: I0312 00:10:19.207362 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4fn9j"] Mar 12 00:10:19 crc kubenswrapper[4948]: I0312 00:10:19.270668 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:19 crc kubenswrapper[4948]: E0312 00:10:19.270945 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 00:10:19.770935596 +0000 UTC m=+219.226539324 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g4kcb" (UID: "af75ce0a-d6bb-48db-a422-08b1d87e1d3f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:19 crc kubenswrapper[4948]: I0312 00:10:19.372064 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 00:10:19 crc kubenswrapper[4948]: E0312 00:10:19.372907 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 00:10:19.87289182 +0000 UTC m=+219.328495558 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:19 crc kubenswrapper[4948]: I0312 00:10:19.428003 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-879f6c89f-9tpt4" podStartSLOduration=180.427985558 podStartE2EDuration="3m0.427985558s" podCreationTimestamp="2026-03-12 00:07:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:10:19.421211589 +0000 UTC m=+218.876815327" watchObservedRunningTime="2026-03-12 00:10:19.427985558 +0000 UTC m=+218.883589296" Mar 12 00:10:19 crc kubenswrapper[4948]: I0312 00:10:19.474199 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:19 crc kubenswrapper[4948]: I0312 00:10:19.474573 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6m4k7" event={"ID":"d08b0c90-3166-4919-88b3-345f5be3efe7","Type":"ContainerStarted","Data":"ffcd4bebd3433d8aa7d0354099cd058c33e0b5f455c9b3095c3cce802ef49acc"} Mar 12 00:10:19 crc kubenswrapper[4948]: E0312 00:10:19.474859 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 00:10:19.974847004 +0000 UTC m=+219.430450742 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g4kcb" (UID: "af75ce0a-d6bb-48db-a422-08b1d87e1d3f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:19 crc kubenswrapper[4948]: W0312 00:10:19.478530 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod2869c0eb_c1c6_4db7_9cf5_17e3300fa14d.slice/crio-6e2f5a464c4cb2ed33ccd478c35f2f65a030594fbd6ca607a49e976778cafc16 WatchSource:0}: Error finding container 6e2f5a464c4cb2ed33ccd478c35f2f65a030594fbd6ca607a49e976778cafc16: Status 404 returned error can't find the container with id 6e2f5a464c4cb2ed33ccd478c35f2f65a030594fbd6ca607a49e976778cafc16 Mar 12 00:10:19 crc kubenswrapper[4948]: I0312 00:10:19.492656 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-xs8b7" event={"ID":"70b489ae-f74b-4e9f-a5b2-71ec1c1f002c","Type":"ContainerStarted","Data":"2d6a4b65d01555f40fe2a96ecefcb3b79dc830d9ee453bd73c18b1f598c080b4"} Mar 12 00:10:19 crc kubenswrapper[4948]: I0312 00:10:19.493383 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-7vhkr" event={"ID":"e6184ca8-dc6e-4d34-ad0e-87c7107acb82","Type":"ContainerStarted","Data":"86fbb7d2f549a52d6982bc4794f251d2febaed48981931d6094eda7995b5851a"} Mar 12 00:10:19 crc kubenswrapper[4948]: I0312 00:10:19.493934 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-rvhq7" event={"ID":"cadf90b6-df8c-48fa-b878-68f2c7d7e49f","Type":"ContainerStarted","Data":"996e3c3dfb3dc1bd0e6e3f15d4c8daa4448ea36781051d570b256a4b98767286"} Mar 12 00:10:19 crc kubenswrapper[4948]: I0312 00:10:19.494486 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-z7w95" event={"ID":"b265c22f-ab92-48e4-9c3e-db6ea6956f6f","Type":"ContainerStarted","Data":"978cc1c5602348ae353cb8d0f479865614108e4b7467d9d6def76e5e3819a265"} Mar 12 00:10:19 crc kubenswrapper[4948]: I0312 00:10:19.504369 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-operator-74547568cd-9k744"] Mar 12 00:10:19 crc kubenswrapper[4948]: I0312 00:10:19.507839 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-z6v9h" event={"ID":"2b9e45c7-edbb-4cbb-9ee2-ae7efb61fbb3","Type":"ContainerStarted","Data":"5ef681b3c9818fdd9ac1f2d7980a15e48c51f085831546899c869e24d98f9a60"} Mar 12 00:10:19 crc kubenswrapper[4948]: I0312 00:10:19.516984 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h6z8v"] Mar 12 00:10:19 crc kubenswrapper[4948]: I0312 00:10:19.558259 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator/migrator-59844c95c7-srr9m"] Mar 12 00:10:19 crc kubenswrapper[4948]: I0312 00:10:19.583914 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-config-operator/openshift-config-operator-7777fb866f-sg87j" event={"ID":"a12c201c-f0ce-4519-8873-e7ec122e7fec","Type":"ContainerStarted","Data":"d099c57ae1d174a957c3670c174e63691ae9fc68cb04ce8e51130870e74a29d8"} Mar 12 00:10:19 crc kubenswrapper[4948]: I0312 00:10:19.586629 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rfxxc"] Mar 12 00:10:19 crc kubenswrapper[4948]: I0312 00:10:19.591358 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 00:10:19 crc kubenswrapper[4948]: E0312 00:10:19.591859 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 00:10:20.091838638 +0000 UTC m=+219.547442376 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:19 crc kubenswrapper[4948]: I0312 00:10:19.594559 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-config-operator/openshift-config-operator-7777fb866f-sg87j" Mar 12 00:10:19 crc kubenswrapper[4948]: I0312 00:10:19.603253 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kvh9w"] Mar 12 00:10:19 crc kubenswrapper[4948]: I0312 00:10:19.605642 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rp8sv" podStartSLOduration=180.605627751 podStartE2EDuration="3m0.605627751s" podCreationTimestamp="2026-03-12 00:07:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:10:19.590109478 +0000 UTC m=+219.045713216" watchObservedRunningTime="2026-03-12 00:10:19.605627751 +0000 UTC m=+219.061231489" Mar 12 00:10:19 crc kubenswrapper[4948]: I0312 00:10:19.618692 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-d24n2" event={"ID":"09e21136-b8c7-4680-ab71-73a86766a7b8","Type":"ContainerStarted","Data":"969ccf6c8e27051339058bb74f36282069c165224d75ad605c558c161df3a110"} Mar 12 00:10:19 crc kubenswrapper[4948]: I0312 00:10:19.620073 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29554560-slpk9" event={"ID":"2c8b7f32-4f80-4f95-b361-7d4687416711","Type":"ContainerStarted","Data":"6c8849999918cbc4ac96da6f6d31b0a3594724331620293f8feb102cd4fadaf5"} Mar 12 00:10:19 crc kubenswrapper[4948]: I0312 00:10:19.620102 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29554560-slpk9" event={"ID":"2c8b7f32-4f80-4f95-b361-7d4687416711","Type":"ContainerStarted","Data":"fc302ca7ca1a53ca39b77dc45506f5093ecef4f025d778c8e2a56f7563e9a34d"} Mar 12 00:10:19 crc kubenswrapper[4948]: I0312 00:10:19.654603 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-dns/dns-default-prqnt"] Mar 12 00:10:19 crc kubenswrapper[4948]: I0312 00:10:19.654860 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-kjggj" event={"ID":"49fe2194-f940-4e0a-a214-fc6080068163","Type":"ContainerStarted","Data":"740446fb577b01edb89a3ac505f3012c280824c8ca20af396b7807b233f608fb"} Mar 12 00:10:19 crc kubenswrapper[4948]: I0312 00:10:19.656485 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bj72v" event={"ID":"2352fbc0-94ff-4b0c-b499-d16b94c279df","Type":"ContainerStarted","Data":"2bdaeee5f0a0ce237ac119c3bc3518601565c1a4170961f5c16d14aa06e01d6f"} Mar 12 00:10:19 crc kubenswrapper[4948]: I0312 00:10:19.674153 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-qq8gg" event={"ID":"892017c6-01a1-4798-9d23-fff4f44c6932","Type":"ContainerStarted","Data":"9be29c24066e5756ca341b459fdb4af68a436869dfa896a93ce4d33273bac889"} Mar 12 00:10:19 crc kubenswrapper[4948]: I0312 00:10:19.674200 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/console-f9d7485db-qq8gg" event={"ID":"892017c6-01a1-4798-9d23-fff4f44c6932","Type":"ContainerStarted","Data":"f9899905392b74e445a0b531ded8cafeacc6a01329c8efe1651463488288c8f8"} Mar 12 00:10:19 crc kubenswrapper[4948]: I0312 00:10:19.683685 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-ingress-canary/ingress-canary-mdhf9"] Mar 12 00:10:19 crc kubenswrapper[4948]: I0312 00:10:19.683791 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-pzgdk" event={"ID":"5e5a84d1-542e-4b58-b991-611bd3202bcf","Type":"ContainerStarted","Data":"777069edaf7f89323f2560f5caffe09cc10881f187acc4b03a1c269c3423f943"} Mar 12 00:10:19 crc kubenswrapper[4948]: I0312 00:10:19.693676 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:19 crc kubenswrapper[4948]: E0312 00:10:19.693996 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 00:10:20.193983886 +0000 UTC m=+219.649587624 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g4kcb" (UID: "af75ce0a-d6bb-48db-a422-08b1d87e1d3f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:19 crc kubenswrapper[4948]: I0312 00:10:19.699898 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-machine-config-operator/machine-config-controller-84d6567774-lkhdg"] Mar 12 00:10:19 crc kubenswrapper[4948]: I0312 00:10:19.708223 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qc4wh" event={"ID":"a79948db-9342-439c-95ae-128c8bd0524d","Type":"ContainerStarted","Data":"07e4dd014f000a07f270e21e37ddf38ac977714822d1247d072d309509705a87"} Mar 12 00:10:19 crc kubenswrapper[4948]: I0312 00:10:19.711706 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console/downloads-7954f5f757-xdh6w" event={"ID":"74ab81d8-55e8-47da-adb5-1f616a100a31","Type":"ContainerStarted","Data":"e7f8bf31bfd2dcdedd7e0fadd4ced8721e3de21b7e3420068c84116596e87767"} Mar 12 00:10:19 crc kubenswrapper[4948]: I0312 00:10:19.712007 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/downloads-7954f5f757-xdh6w" Mar 12 00:10:19 crc kubenswrapper[4948]: I0312 00:10:19.712798 4948 patch_prober.go:28] interesting pod/downloads-7954f5f757-xdh6w container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" start-of-body= Mar 12 00:10:19 crc kubenswrapper[4948]: I0312 00:10:19.712830 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-xdh6w" podUID="74ab81d8-55e8-47da-adb5-1f616a100a31" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" Mar 12 00:10:19 crc kubenswrapper[4948]: I0312 00:10:19.754200 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-n4545" event={"ID":"a6da8984-cc2d-4b58-b6da-d7fc2d1e1725","Type":"ContainerStarted","Data":"f89636361b2628cc1dbcdd6431bc6726c676431f1ce68405e7840660b13320c6"} Mar 12 00:10:19 crc kubenswrapper[4948]: I0312 00:10:19.762168 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication-operator/authentication-operator-69f744f599-5cwjt" podStartSLOduration=180.76215237 podStartE2EDuration="3m0.76215237s" podCreationTimestamp="2026-03-12 00:07:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:10:19.761082845 +0000 UTC m=+219.216686583" watchObservedRunningTime="2026-03-12 00:10:19.76215237 +0000 UTC m=+219.217756108" Mar 12 00:10:19 crc kubenswrapper[4948]: I0312 00:10:19.794295 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-qsf8x" event={"ID":"f90934d6-9305-4513-8d84-cd6a6fcc4a59","Type":"ContainerStarted","Data":"56f73e2a84df78402185d6ddeb2c483aaed0eb93cd03771074df5b28ef703b8a"} Mar 12 00:10:19 crc kubenswrapper[4948]: I0312 00:10:19.794351 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-558db77b4-qsf8x" Mar 12 00:10:19 crc kubenswrapper[4948]: I0312 00:10:19.797272 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 00:10:19 crc kubenswrapper[4948]: E0312 00:10:19.804047 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 00:10:20.304032309 +0000 UTC m=+219.759636047 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:19 crc kubenswrapper[4948]: I0312 00:10:19.808145 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rp8sv" Mar 12 00:10:19 crc kubenswrapper[4948]: I0312 00:10:19.816493 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-879f6c89f-9tpt4" Mar 12 00:10:19 crc kubenswrapper[4948]: I0312 00:10:19.892833 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jlrln"] Mar 12 00:10:19 crc kubenswrapper[4948]: I0312 00:10:19.948402 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:19 crc kubenswrapper[4948]: E0312 00:10:19.952151 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 00:10:20.452135002 +0000 UTC m=+219.907738740 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g4kcb" (UID: "af75ce0a-d6bb-48db-a422-08b1d87e1d3f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:20 crc kubenswrapper[4948]: I0312 00:10:20.014941 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/machine-api-operator-5694c8668f-g7g94" podStartSLOduration=181.014926649 podStartE2EDuration="3m1.014926649s" podCreationTimestamp="2026-03-12 00:07:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:10:20.012738368 +0000 UTC m=+219.468342106" watchObservedRunningTime="2026-03-12 00:10:20.014926649 +0000 UTC m=+219.470530387" Mar 12 00:10:20 crc kubenswrapper[4948]: I0312 00:10:20.066558 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 00:10:20 crc kubenswrapper[4948]: E0312 00:10:20.067395 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 00:10:20.567378166 +0000 UTC m=+220.022981904 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:20 crc kubenswrapper[4948]: I0312 00:10:20.103757 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29554570-2b2hb"] Mar 12 00:10:20 crc kubenswrapper[4948]: I0312 00:10:20.173520 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:20 crc kubenswrapper[4948]: E0312 00:10:20.173855 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 00:10:20.673843854 +0000 UTC m=+220.129447592 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g4kcb" (UID: "af75ce0a-d6bb-48db-a422-08b1d87e1d3f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:20 crc kubenswrapper[4948]: I0312 00:10:20.243860 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-config-operator/openshift-config-operator-7777fb866f-sg87j" podStartSLOduration=181.243837571 podStartE2EDuration="3m1.243837571s" podCreationTimestamp="2026-03-12 00:07:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:10:20.23696854 +0000 UTC m=+219.692572278" watchObservedRunningTime="2026-03-12 00:10:20.243837571 +0000 UTC m=+219.699441309" Mar 12 00:10:20 crc kubenswrapper[4948]: I0312 00:10:20.259475 4948 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 12 00:10:20 crc kubenswrapper[4948]: I0312 00:10:20.269612 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-pruner-29554560-slpk9" podStartSLOduration=181.269592732 podStartE2EDuration="3m1.269592732s" podCreationTimestamp="2026-03-12 00:07:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:10:20.263443219 +0000 UTC m=+219.719046957" watchObservedRunningTime="2026-03-12 00:10:20.269592732 +0000 UTC m=+219.725196470" Mar 12 00:10:20 crc kubenswrapper[4948]: I0312 00:10:20.274440 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 00:10:20 crc kubenswrapper[4948]: E0312 00:10:20.274662 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 00:10:20.774648121 +0000 UTC m=+220.230251859 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:20 crc kubenswrapper[4948]: I0312 00:10:20.299168 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/console-f9d7485db-qq8gg" podStartSLOduration=181.299151523 podStartE2EDuration="3m1.299151523s" podCreationTimestamp="2026-03-12 00:07:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:10:20.285015063 +0000 UTC m=+219.740618801" watchObservedRunningTime="2026-03-12 00:10:20.299151523 +0000 UTC m=+219.754755261" Mar 12 00:10:20 crc kubenswrapper[4948]: I0312 00:10:20.300339 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-service-ca/service-ca-9c57cc56f-q79m4"] Mar 12 00:10:20 crc kubenswrapper[4948]: I0312 00:10:20.330940 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["hostpath-provisioner/csi-hostpathplugin-5q9rs"] Mar 12 00:10:20 crc kubenswrapper[4948]: I0312 00:10:20.383992 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager-operator/openshift-controller-manager-operator-756b6f6bc6-qc4wh" podStartSLOduration=181.383977547 podStartE2EDuration="3m1.383977547s" podCreationTimestamp="2026-03-12 00:07:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:10:20.340570272 +0000 UTC m=+219.796174010" watchObservedRunningTime="2026-03-12 00:10:20.383977547 +0000 UTC m=+219.839581285" Mar 12 00:10:20 crc kubenswrapper[4948]: I0312 00:10:20.386784 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:20 crc kubenswrapper[4948]: E0312 00:10:20.387210 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 00:10:20.887194572 +0000 UTC m=+220.342798310 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g4kcb" (UID: "af75ce0a-d6bb-48db-a422-08b1d87e1d3f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:20 crc kubenswrapper[4948]: I0312 00:10:20.401421 4948 ???:1] "http: TLS handshake error from 192.168.126.11:35488: no serving certificate available for the kubelet" Mar 12 00:10:20 crc kubenswrapper[4948]: I0312 00:10:20.401625 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-558db77b4-qsf8x" Mar 12 00:10:20 crc kubenswrapper[4948]: I0312 00:10:20.431076 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-558db77b4-qsf8x" podStartSLOduration=181.431058368 podStartE2EDuration="3m1.431058368s" podCreationTimestamp="2026-03-12 00:07:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:10:20.425289372 +0000 UTC m=+219.880893110" watchObservedRunningTime="2026-03-12 00:10:20.431058368 +0000 UTC m=+219.886662106" Mar 12 00:10:20 crc kubenswrapper[4948]: I0312 00:10:20.463550 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-machine-approver/machine-approver-56656f9798-n4545" podStartSLOduration=181.463530747 podStartE2EDuration="3m1.463530747s" podCreationTimestamp="2026-03-12 00:07:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:10:20.459330579 +0000 UTC m=+219.914934317" watchObservedRunningTime="2026-03-12 00:10:20.463530747 +0000 UTC m=+219.919134485" Mar 12 00:10:20 crc kubenswrapper[4948]: I0312 00:10:20.468473 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29554560-8mgv9"] Mar 12 00:10:20 crc kubenswrapper[4948]: I0312 00:10:20.486707 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-v59sv"] Mar 12 00:10:20 crc kubenswrapper[4948]: I0312 00:10:20.487349 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 00:10:20 crc kubenswrapper[4948]: E0312 00:10:20.487657 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 00:10:20.98764398 +0000 UTC m=+220.443247718 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:20 crc kubenswrapper[4948]: I0312 00:10:20.503354 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5qjdn"] Mar 12 00:10:20 crc kubenswrapper[4948]: I0312 00:10:20.503724 4948 ???:1] "http: TLS handshake error from 192.168.126.11:35498: no serving certificate available for the kubelet" Mar 12 00:10:20 crc kubenswrapper[4948]: I0312 00:10:20.509897 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console/downloads-7954f5f757-xdh6w" podStartSLOduration=181.50988237 podStartE2EDuration="3m1.50988237s" podCreationTimestamp="2026-03-12 00:07:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:10:20.509710326 +0000 UTC m=+219.965314064" watchObservedRunningTime="2026-03-12 00:10:20.50988237 +0000 UTC m=+219.965486108" Mar 12 00:10:20 crc kubenswrapper[4948]: I0312 00:10:20.519468 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dc6wb"] Mar 12 00:10:20 crc kubenswrapper[4948]: W0312 00:10:20.536940 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4ba12474_c71f_429c_bcdf_9d74eb0b4b7a.slice/crio-ecbd919c13b434ed7936bb218731912c7d3b7f919c3ac3a995654c91e36965a4 WatchSource:0}: Error finding container ecbd919c13b434ed7936bb218731912c7d3b7f919c3ac3a995654c91e36965a4: Status 404 returned error can't find the container with id ecbd919c13b434ed7936bb218731912c7d3b7f919c3ac3a995654c91e36965a4 Mar 12 00:10:20 crc kubenswrapper[4948]: I0312 00:10:20.588594 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:20 crc kubenswrapper[4948]: E0312 00:10:20.588891 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 00:10:21.088880217 +0000 UTC m=+220.544483955 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g4kcb" (UID: "af75ce0a-d6bb-48db-a422-08b1d87e1d3f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:20 crc kubenswrapper[4948]: I0312 00:10:20.599673 4948 ???:1] "http: TLS handshake error from 192.168.126.11:35502: no serving certificate available for the kubelet" Mar 12 00:10:20 crc kubenswrapper[4948]: I0312 00:10:20.690360 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 00:10:20 crc kubenswrapper[4948]: E0312 00:10:20.690911 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 00:10:21.190895712 +0000 UTC m=+220.646499450 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:20 crc kubenswrapper[4948]: I0312 00:10:20.704814 4948 ???:1] "http: TLS handshake error from 192.168.126.11:35508: no serving certificate available for the kubelet" Mar 12 00:10:20 crc kubenswrapper[4948]: I0312 00:10:20.801488 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:20 crc kubenswrapper[4948]: E0312 00:10:20.801877 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 00:10:21.301861335 +0000 UTC m=+220.757465073 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g4kcb" (UID: "af75ce0a-d6bb-48db-a422-08b1d87e1d3f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:20 crc kubenswrapper[4948]: I0312 00:10:20.804442 4948 ???:1] "http: TLS handshake error from 192.168.126.11:35510: no serving certificate available for the kubelet" Mar 12 00:10:20 crc kubenswrapper[4948]: I0312 00:10:20.812354 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-zkbhv" event={"ID":"db40ba09-a72f-43d5-8037-e71c96e2bf5b","Type":"ContainerStarted","Data":"77256205eaf26927b82537a261f9e9f9e55832d90d862711c36f6862edd98955"} Mar 12 00:10:20 crc kubenswrapper[4948]: I0312 00:10:20.812396 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-zkbhv" event={"ID":"db40ba09-a72f-43d5-8037-e71c96e2bf5b","Type":"ContainerStarted","Data":"d150043fc21742d2456851014afeb8e8fe132595929076230128b9b80cd74e7a"} Mar 12 00:10:20 crc kubenswrapper[4948]: I0312 00:10:20.813858 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-zkbhv" Mar 12 00:10:20 crc kubenswrapper[4948]: I0312 00:10:20.819415 4948 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-zkbhv container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.40:8080/healthz\": dial tcp 10.217.0.40:8080: connect: connection refused" start-of-body= Mar 12 00:10:20 crc kubenswrapper[4948]: I0312 00:10:20.819470 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-zkbhv" podUID="db40ba09-a72f-43d5-8037-e71c96e2bf5b" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.40:8080/healthz\": dial tcp 10.217.0.40:8080: connect: connection refused" Mar 12 00:10:20 crc kubenswrapper[4948]: I0312 00:10:20.820399 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-mdhf9" event={"ID":"15828966-e769-43d4-b4a3-7a509fa72814","Type":"ContainerStarted","Data":"30e9d3c4994c6e399c7f06b5dcd7087981cea78cbbd4d853104004b49084bc25"} Mar 12 00:10:20 crc kubenswrapper[4948]: I0312 00:10:20.836491 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns-operator/dns-operator-744455d44c-xs8b7" event={"ID":"70b489ae-f74b-4e9f-a5b2-71ec1c1f002c","Type":"ContainerStarted","Data":"4f8b2a696c377fcfb04f0c7801e8802d127a60a494214ff6fe1450af13609fd1"} Mar 12 00:10:20 crc kubenswrapper[4948]: I0312 00:10:20.855576 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-server-rvhq7" event={"ID":"cadf90b6-df8c-48fa-b878-68f2c7d7e49f","Type":"ContainerStarted","Data":"daf9fb507178c4673a02cc64120ca8e2136f84fc5075e8d804737e475eb16065"} Mar 12 00:10:20 crc kubenswrapper[4948]: I0312 00:10:20.871114 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-zkbhv" podStartSLOduration=181.871100815 podStartE2EDuration="3m1.871100815s" podCreationTimestamp="2026-03-12 00:07:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:10:20.870716806 +0000 UTC m=+220.326320544" watchObservedRunningTime="2026-03-12 00:10:20.871100815 +0000 UTC m=+220.326704553" Mar 12 00:10:20 crc kubenswrapper[4948]: I0312 00:10:20.885696 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-5q9rs" event={"ID":"4ba12474-c71f-429c-bcdf-9d74eb0b4b7a","Type":"ContainerStarted","Data":"ecbd919c13b434ed7936bb218731912c7d3b7f919c3ac3a995654c91e36965a4"} Mar 12 00:10:20 crc kubenswrapper[4948]: I0312 00:10:20.900074 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5qjdn" event={"ID":"3e538a31-98b5-4353-8038-4a7580be89dd","Type":"ContainerStarted","Data":"49ffa1f39073f8705d300c9e2f97bb07b5093edc9b9880956a1a2e0dc5aff0d7"} Mar 12 00:10:20 crc kubenswrapper[4948]: I0312 00:10:20.903940 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 00:10:20 crc kubenswrapper[4948]: E0312 00:10:20.910231 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 00:10:21.410213959 +0000 UTC m=+220.865817697 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:20 crc kubenswrapper[4948]: I0312 00:10:20.915847 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns-operator/dns-operator-744455d44c-xs8b7" podStartSLOduration=181.91583305 podStartE2EDuration="3m1.91583305s" podCreationTimestamp="2026-03-12 00:07:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:10:20.913847163 +0000 UTC m=+220.369450901" watchObservedRunningTime="2026-03-12 00:10:20.91583305 +0000 UTC m=+220.371436788" Mar 12 00:10:20 crc kubenswrapper[4948]: I0312 00:10:20.981142 4948 ???:1] "http: TLS handshake error from 192.168.126.11:35516: no serving certificate available for the kubelet" Mar 12 00:10:20 crc kubenswrapper[4948]: I0312 00:10:20.992719 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lkhdg" event={"ID":"bd97cad6-0d6d-4bec-9d2f-e17b03b220f6","Type":"ContainerStarted","Data":"5066e470e2bab1b6bff3238220cf64aeb20a25707f8317afbfd33899a913d7f7"} Mar 12 00:10:21 crc kubenswrapper[4948]: I0312 00:10:21.006868 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:21 crc kubenswrapper[4948]: E0312 00:10:21.007488 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 00:10:21.507472972 +0000 UTC m=+220.963076710 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g4kcb" (UID: "af75ce0a-d6bb-48db-a422-08b1d87e1d3f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:21 crc kubenswrapper[4948]: I0312 00:10:21.029869 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fx7lk" event={"ID":"f980811d-53fd-4c64-96c5-19283afcbb1f","Type":"ContainerStarted","Data":"64f34aa75730b4f1854de495aefa434103324ab202373ed77ba9656b5da702e6"} Mar 12 00:10:21 crc kubenswrapper[4948]: I0312 00:10:21.032664 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29554560-8mgv9" event={"ID":"5b801839-e39a-48a9-9eee-1bb95df5fbfb","Type":"ContainerStarted","Data":"a043eedcdc4b607d199f5e715c5a48cdc1eac9b39adc704584885494a0fa2ab7"} Mar 12 00:10:21 crc kubenswrapper[4948]: I0312 00:10:21.054615 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-z6v9h" event={"ID":"2b9e45c7-edbb-4cbb-9ee2-ae7efb61fbb3","Type":"ContainerStarted","Data":"fba0e5a28851e43b14ffae7bfcc086d7a648e7f72a1be39a7f28284f7f2e020a"} Mar 12 00:10:21 crc kubenswrapper[4948]: I0312 00:10:21.067929 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dc6wb" event={"ID":"7599787e-ea15-40ca-b365-fdf406e3c337","Type":"ContainerStarted","Data":"9251b5ff5f8c9d4e41830cbe41ee4a843f87264858f947f216524b1d53685254"} Mar 12 00:10:21 crc kubenswrapper[4948]: I0312 00:10:21.078430 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-v59sv" event={"ID":"7d1dc3d3-060c-4aba-aa54-0a3ea767f299","Type":"ContainerStarted","Data":"a4041f33016e3723867188ee2f5bcda0fc19ed37197305b0214c3e999622d4bc"} Mar 12 00:10:21 crc kubenswrapper[4948]: I0312 00:10:21.091572 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-server-rvhq7" podStartSLOduration=6.091554648 podStartE2EDuration="6.091554648s" podCreationTimestamp="2026-03-12 00:10:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:10:20.958781484 +0000 UTC m=+220.414385222" watchObservedRunningTime="2026-03-12 00:10:21.091554648 +0000 UTC m=+220.547158386" Mar 12 00:10:21 crc kubenswrapper[4948]: I0312 00:10:21.106573 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bj72v" event={"ID":"2352fbc0-94ff-4b0c-b499-d16b94c279df","Type":"ContainerStarted","Data":"d5bcef23ae5561f79ef384353413efb2fc7b7ba42748a95371c889b8d655075e"} Mar 12 00:10:21 crc kubenswrapper[4948]: I0312 00:10:21.107873 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 00:10:21 crc kubenswrapper[4948]: E0312 00:10:21.108733 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 00:10:21.60871494 +0000 UTC m=+221.064318678 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:21 crc kubenswrapper[4948]: I0312 00:10:21.139226 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress/router-default-5444994796-7vhkr" event={"ID":"e6184ca8-dc6e-4d34-ad0e-87c7107acb82","Type":"ContainerStarted","Data":"9c2c184af4dbf9a59aa46a4b44e3fb91fe0c17adb4422b5112c8c6c464bfdbab"} Mar 12 00:10:21 crc kubenswrapper[4948]: I0312 00:10:21.150278 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fx7lk" podStartSLOduration=182.15026017 podStartE2EDuration="3m2.15026017s" podCreationTimestamp="2026-03-12 00:07:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:10:21.092148472 +0000 UTC m=+220.547752210" watchObservedRunningTime="2026-03-12 00:10:21.15026017 +0000 UTC m=+220.605863908" Mar 12 00:10:21 crc kubenswrapper[4948]: I0312 00:10:21.150438 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-api/control-plane-machine-set-operator-78cbb6b69f-bj72v" podStartSLOduration=182.150433224 podStartE2EDuration="3m2.150433224s" podCreationTimestamp="2026-03-12 00:07:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:10:21.148355156 +0000 UTC m=+220.603958894" watchObservedRunningTime="2026-03-12 00:10:21.150433224 +0000 UTC m=+220.606036972" Mar 12 00:10:21 crc kubenswrapper[4948]: I0312 00:10:21.206238 4948 ???:1] "http: TLS handshake error from 192.168.126.11:35520: no serving certificate available for the kubelet" Mar 12 00:10:21 crc kubenswrapper[4948]: I0312 00:10:21.209038 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:21 crc kubenswrapper[4948]: E0312 00:10:21.210017 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 00:10:21.710005167 +0000 UTC m=+221.165608905 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g4kcb" (UID: "af75ce0a-d6bb-48db-a422-08b1d87e1d3f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:21 crc kubenswrapper[4948]: I0312 00:10:21.214618 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9k744" event={"ID":"1196fdb1-9c09-4982-9c37-35f46fabc1c5","Type":"ContainerStarted","Data":"b83c75c25518d5cf3b01e3c802a60ecac156cd2350983e6b25dd90a185881a16"} Mar 12 00:10:21 crc kubenswrapper[4948]: I0312 00:10:21.236343 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29554570-2b2hb" event={"ID":"2fb324d6-44d4-440d-ae60-a5356a3a057f","Type":"ContainerStarted","Data":"1d4d414e0f4148242881bbd99b11885b2652663b62baaf55ce4bb47096bd3a56"} Mar 12 00:10:21 crc kubenswrapper[4948]: I0312 00:10:21.268443 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-q79m4" event={"ID":"141bb112-1f07-4db5-9a4e-b7d94f12c5e2","Type":"ContainerStarted","Data":"e136a82059d74dcf12a5077bee732ef6fc758e11111ced967f033e344d17610d"} Mar 12 00:10:21 crc kubenswrapper[4948]: I0312 00:10:21.291960 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rfxxc" event={"ID":"29cf53b1-0215-4dd4-a19f-24a48d4ded9d","Type":"ContainerStarted","Data":"d2f15a7fd5997850d8ba4675f407f7c07ac132a3b7f0e75bd15c0a5ce0c320c6"} Mar 12 00:10:21 crc kubenswrapper[4948]: I0312 00:10:21.315580 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 00:10:21 crc kubenswrapper[4948]: E0312 00:10:21.315964 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 00:10:21.815947894 +0000 UTC m=+221.271551632 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:21 crc kubenswrapper[4948]: I0312 00:10:21.332365 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4fn9j" event={"ID":"2869c0eb-c1c6-4db7-9cf5-17e3300fa14d","Type":"ContainerStarted","Data":"be9883442737f44ae14bf95af8c28d28fbfdfa8c36464c94996f414898a4e43e"} Mar 12 00:10:21 crc kubenswrapper[4948]: I0312 00:10:21.332513 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4fn9j" event={"ID":"2869c0eb-c1c6-4db7-9cf5-17e3300fa14d","Type":"ContainerStarted","Data":"6e2f5a464c4cb2ed33ccd478c35f2f65a030594fbd6ca607a49e976778cafc16"} Mar 12 00:10:21 crc kubenswrapper[4948]: I0312 00:10:21.378550 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress/router-default-5444994796-7vhkr" podStartSLOduration=182.378533017 podStartE2EDuration="3m2.378533017s" podCreationTimestamp="2026-03-12 00:07:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:10:21.183901237 +0000 UTC m=+220.639504975" watchObservedRunningTime="2026-03-12 00:10:21.378533017 +0000 UTC m=+220.834136755" Mar 12 00:10:21 crc kubenswrapper[4948]: I0312 00:10:21.388140 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kvh9w" event={"ID":"1666a214-546a-4063-a627-e990781797c8","Type":"ContainerStarted","Data":"4e8f1adf97fdb07dde1fa537795299beeae39fa070e13a25032c32736be1fdbb"} Mar 12 00:10:21 crc kubenswrapper[4948]: I0312 00:10:21.396307 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca-operator/service-ca-operator-777779d784-kjggj" event={"ID":"49fe2194-f940-4e0a-a214-fc6080068163","Type":"ContainerStarted","Data":"688c1630d103e89d69cfd986270f27959cbc647ad455b7fe32f69a44be64cdcc"} Mar 12 00:10:21 crc kubenswrapper[4948]: I0312 00:10:21.412046 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gc7vl" event={"ID":"4dc52af7-648a-40e7-bdae-50672f5143dc","Type":"ContainerStarted","Data":"57052511f6219077bdf8bd0c061bfdd0a71f25d63eb652dc076731391e2aafd4"} Mar 12 00:10:21 crc kubenswrapper[4948]: I0312 00:10:21.416951 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:21 crc kubenswrapper[4948]: E0312 00:10:21.418044 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 00:10:21.91803329 +0000 UTC m=+221.373637028 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g4kcb" (UID: "af75ce0a-d6bb-48db-a422-08b1d87e1d3f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:21 crc kubenswrapper[4948]: I0312 00:10:21.448137 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-6jfc4" event={"ID":"2040ed8d-ae32-4909-9161-bc7ffbe49edd","Type":"ContainerStarted","Data":"6aac0fe3debebe2863f9c8b0984b4680a9ef9f14840cc637a7686fb9b4b1ee63"} Mar 12 00:10:21 crc kubenswrapper[4948]: I0312 00:10:21.474365 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-etcd-operator/etcd-operator-b45778765-pzgdk" event={"ID":"5e5a84d1-542e-4b58-b991-611bd3202bcf","Type":"ContainerStarted","Data":"4bb7ee40b243b6debf0ddd0113e7827542527d974230dfcd900d7ab840b42cdb"} Mar 12 00:10:21 crc kubenswrapper[4948]: I0312 00:10:21.494609 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-prqnt" event={"ID":"2851d901-f861-432b-a61e-5e0391f1b2b0","Type":"ContainerStarted","Data":"b4207b23d3eee9cec78e48ca1451f4da4cbb891ccfeaa5f7f1c18e989ec90a4c"} Mar 12 00:10:21 crc kubenswrapper[4948]: I0312 00:10:21.517875 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 00:10:21 crc kubenswrapper[4948]: E0312 00:10:21.518906 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 00:10:22.018863217 +0000 UTC m=+221.474466945 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:21 crc kubenswrapper[4948]: I0312 00:10:21.582805 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-console-operator/console-operator-58897d9998-z7w95" event={"ID":"b265c22f-ab92-48e4-9c3e-db6ea6956f6f","Type":"ContainerStarted","Data":"586834ff64b3b6dc91adfe84c65c1e074372d9eac4f44a317c96c9cd94798936"} Mar 12 00:10:21 crc kubenswrapper[4948]: I0312 00:10:21.583859 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console-operator/console-operator-58897d9998-z7w95" Mar 12 00:10:21 crc kubenswrapper[4948]: I0312 00:10:21.605206 4948 ???:1] "http: TLS handshake error from 192.168.126.11:35528: no serving certificate available for the kubelet" Mar 12 00:10:21 crc kubenswrapper[4948]: I0312 00:10:21.616838 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-srr9m" event={"ID":"677936de-2686-46cf-99e9-cae2355c42b4","Type":"ContainerStarted","Data":"f5a738c9b1f143aea98bb6ea49f0411c9b78bbf6bb5dff841bd0b066ceb23e18"} Mar 12 00:10:21 crc kubenswrapper[4948]: I0312 00:10:21.619062 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:21 crc kubenswrapper[4948]: E0312 00:10:21.619378 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 00:10:22.119366137 +0000 UTC m=+221.574969875 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g4kcb" (UID: "af75ce0a-d6bb-48db-a422-08b1d87e1d3f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:21 crc kubenswrapper[4948]: I0312 00:10:21.658527 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jlrln" event={"ID":"d33d5db8-961e-4c7a-950e-cb3f88a3f9b0","Type":"ContainerStarted","Data":"e54fa4f00d22c65575d6a96104e0ab9b2363fb069135521e974a906e2fe0f9cc"} Mar 12 00:10:21 crc kubenswrapper[4948]: I0312 00:10:21.674547 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h6z8v" event={"ID":"12c99e25-86bd-4923-ab44-55e571661c0f","Type":"ContainerStarted","Data":"49c2224995f61fd7d2c753d591d12c6cd9bf7dca58d9b01e7664b591d78aa29b"} Mar 12 00:10:21 crc kubenswrapper[4948]: I0312 00:10:21.695905 4948 patch_prober.go:28] interesting pod/downloads-7954f5f757-xdh6w container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" start-of-body= Mar 12 00:10:21 crc kubenswrapper[4948]: I0312 00:10:21.695952 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-xdh6w" podUID="74ab81d8-55e8-47da-adb5-1f616a100a31" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" Mar 12 00:10:21 crc kubenswrapper[4948]: I0312 00:10:21.703457 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-9tpt4"] Mar 12 00:10:21 crc kubenswrapper[4948]: I0312 00:10:21.721594 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 00:10:21 crc kubenswrapper[4948]: E0312 00:10:21.722559 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 00:10:22.222538229 +0000 UTC m=+221.678141967 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:21 crc kubenswrapper[4948]: I0312 00:10:21.789238 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-rp8sv"] Mar 12 00:10:21 crc kubenswrapper[4948]: I0312 00:10:21.823561 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:21 crc kubenswrapper[4948]: E0312 00:10:21.823935 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 00:10:22.323923639 +0000 UTC m=+221.779527377 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g4kcb" (UID: "af75ce0a-d6bb-48db-a422-08b1d87e1d3f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:21 crc kubenswrapper[4948]: I0312 00:10:21.885564 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-config-operator/openshift-config-operator-7777fb866f-sg87j" Mar 12 00:10:21 crc kubenswrapper[4948]: I0312 00:10:21.924508 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h6z8v" podStartSLOduration=182.92449362 podStartE2EDuration="3m2.92449362s" podCreationTimestamp="2026-03-12 00:07:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:10:21.895041952 +0000 UTC m=+221.350645690" watchObservedRunningTime="2026-03-12 00:10:21.92449362 +0000 UTC m=+221.380097358" Mar 12 00:10:21 crc kubenswrapper[4948]: I0312 00:10:21.924775 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 00:10:21 crc kubenswrapper[4948]: E0312 00:10:21.925206 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 00:10:22.425189727 +0000 UTC m=+221.880793465 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:21 crc kubenswrapper[4948]: I0312 00:10:21.925351 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kvh9w" podStartSLOduration=182.92534446 podStartE2EDuration="3m2.92534446s" podCreationTimestamp="2026-03-12 00:07:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:10:21.9223458 +0000 UTC m=+221.377949538" watchObservedRunningTime="2026-03-12 00:10:21.92534446 +0000 UTC m=+221.380948198" Mar 12 00:10:21 crc kubenswrapper[4948]: I0312 00:10:21.952103 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-etcd-operator/etcd-operator-b45778765-pzgdk" podStartSLOduration=182.952085805 podStartE2EDuration="3m2.952085805s" podCreationTimestamp="2026-03-12 00:07:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:10:21.950719543 +0000 UTC m=+221.406323281" watchObservedRunningTime="2026-03-12 00:10:21.952085805 +0000 UTC m=+221.407689543" Mar 12 00:10:21 crc kubenswrapper[4948]: I0312 00:10:21.973013 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-console-operator/console-operator-58897d9998-z7w95" podStartSLOduration=182.972997604 podStartE2EDuration="3m2.972997604s" podCreationTimestamp="2026-03-12 00:07:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:10:21.97070228 +0000 UTC m=+221.426306018" watchObservedRunningTime="2026-03-12 00:10:21.972997604 +0000 UTC m=+221.428601342" Mar 12 00:10:21 crc kubenswrapper[4948]: I0312 00:10:21.986078 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gc7vl" podStartSLOduration=182.986058379 podStartE2EDuration="3m2.986058379s" podCreationTimestamp="2026-03-12 00:07:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:10:21.983600782 +0000 UTC m=+221.439204520" watchObservedRunningTime="2026-03-12 00:10:21.986058379 +0000 UTC m=+221.441662117" Mar 12 00:10:22 crc kubenswrapper[4948]: I0312 00:10:22.013664 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca-operator/service-ca-operator-777779d784-kjggj" podStartSLOduration=183.013650515 podStartE2EDuration="3m3.013650515s" podCreationTimestamp="2026-03-12 00:07:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:10:22.012351704 +0000 UTC m=+221.467955442" watchObservedRunningTime="2026-03-12 00:10:22.013650515 +0000 UTC m=+221.469254243" Mar 12 00:10:22 crc kubenswrapper[4948]: I0312 00:10:22.027021 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:22 crc kubenswrapper[4948]: E0312 00:10:22.027410 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 00:10:22.527386756 +0000 UTC m=+221.982990494 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g4kcb" (UID: "af75ce0a-d6bb-48db-a422-08b1d87e1d3f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:22 crc kubenswrapper[4948]: I0312 00:10:22.033564 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/cluster-image-registry-operator-dc59b4c8b-4fn9j" podStartSLOduration=183.03354881 podStartE2EDuration="3m3.03354881s" podCreationTimestamp="2026-03-12 00:07:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:10:22.030683613 +0000 UTC m=+221.486287361" watchObservedRunningTime="2026-03-12 00:10:22.03354881 +0000 UTC m=+221.489152548" Mar 12 00:10:22 crc kubenswrapper[4948]: I0312 00:10:22.094807 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-ingress/router-default-5444994796-7vhkr" Mar 12 00:10:22 crc kubenswrapper[4948]: I0312 00:10:22.101707 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-apiserver/apiserver-76f77b778f-6jfc4" podStartSLOduration=183.101691073 podStartE2EDuration="3m3.101691073s" podCreationTimestamp="2026-03-12 00:07:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:10:22.078701095 +0000 UTC m=+221.534304833" watchObservedRunningTime="2026-03-12 00:10:22.101691073 +0000 UTC m=+221.557294811" Mar 12 00:10:22 crc kubenswrapper[4948]: I0312 00:10:22.102101 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-srr9m" podStartSLOduration=183.102095142 podStartE2EDuration="3m3.102095142s" podCreationTimestamp="2026-03-12 00:07:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:10:22.099641095 +0000 UTC m=+221.555244833" watchObservedRunningTime="2026-03-12 00:10:22.102095142 +0000 UTC m=+221.557698880" Mar 12 00:10:22 crc kubenswrapper[4948]: I0312 00:10:22.102818 4948 patch_prober.go:28] interesting pod/router-default-5444994796-7vhkr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 12 00:10:22 crc kubenswrapper[4948]: [-]has-synced failed: reason withheld Mar 12 00:10:22 crc kubenswrapper[4948]: [+]process-running ok Mar 12 00:10:22 crc kubenswrapper[4948]: healthz check failed Mar 12 00:10:22 crc kubenswrapper[4948]: I0312 00:10:22.102911 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7vhkr" podUID="e6184ca8-dc6e-4d34-ad0e-87c7107acb82" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 12 00:10:22 crc kubenswrapper[4948]: I0312 00:10:22.127810 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 00:10:22 crc kubenswrapper[4948]: E0312 00:10:22.128174 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 00:10:22.628159402 +0000 UTC m=+222.083763140 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:22 crc kubenswrapper[4948]: I0312 00:10:22.229076 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:10:22 crc kubenswrapper[4948]: I0312 00:10:22.229129 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:10:22 crc kubenswrapper[4948]: I0312 00:10:22.229162 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:22 crc kubenswrapper[4948]: I0312 00:10:22.229181 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:10:22 crc kubenswrapper[4948]: I0312 00:10:22.229216 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:10:22 crc kubenswrapper[4948]: I0312 00:10:22.230853 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"nginx-conf\" (UniqueName: \"kubernetes.io/configmap/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-nginx-conf\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:10:22 crc kubenswrapper[4948]: E0312 00:10:22.231262 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 00:10:22.731242871 +0000 UTC m=+222.186846599 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g4kcb" (UID: "af75ce0a-d6bb-48db-a422-08b1d87e1d3f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:22 crc kubenswrapper[4948]: I0312 00:10:22.236429 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fx7lk" Mar 12 00:10:22 crc kubenswrapper[4948]: I0312 00:10:22.236781 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fx7lk" Mar 12 00:10:22 crc kubenswrapper[4948]: I0312 00:10:22.244506 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s2dwl\" (UniqueName: \"kubernetes.io/projected/9d751cbb-f2e2-430d-9754-c882a5e924a5-kube-api-access-s2dwl\") pod \"network-check-source-55646444c4-trplf\" (UID: \"9d751cbb-f2e2-430d-9754-c882a5e924a5\") " pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:10:22 crc kubenswrapper[4948]: I0312 00:10:22.248169 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cqllr\" (UniqueName: \"kubernetes.io/projected/3b6479f0-333b-4a96-9adf-2099afdc2447-kube-api-access-cqllr\") pod \"network-check-target-xd92c\" (UID: \"3b6479f0-333b-4a96-9adf-2099afdc2447\") " pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:10:22 crc kubenswrapper[4948]: I0312 00:10:22.268617 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fx7lk" Mar 12 00:10:22 crc kubenswrapper[4948]: I0312 00:10:22.276869 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"networking-console-plugin-cert\" (UniqueName: \"kubernetes.io/secret/5fe485a1-e14f-4c09-b5b9-f252bc42b7e8-networking-console-plugin-cert\") pod \"networking-console-plugin-85b44fc459-gdk6g\" (UID: \"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8\") " pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:10:22 crc kubenswrapper[4948]: I0312 00:10:22.330694 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 00:10:22 crc kubenswrapper[4948]: I0312 00:10:22.330830 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/be985ff1-de8a-431f-a230-67894ebe52d5-metrics-certs\") pod \"network-metrics-daemon-mqmtd\" (UID: \"be985ff1-de8a-431f-a230-67894ebe52d5\") " pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:10:22 crc kubenswrapper[4948]: E0312 00:10:22.336677 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 00:10:22.836652426 +0000 UTC m=+222.292256164 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:22 crc kubenswrapper[4948]: I0312 00:10:22.354717 4948 ???:1] "http: TLS handshake error from 192.168.126.11:35534: no serving certificate available for the kubelet" Mar 12 00:10:22 crc kubenswrapper[4948]: I0312 00:10:22.355221 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"metrics-certs\" (UniqueName: \"kubernetes.io/secret/be985ff1-de8a-431f-a230-67894ebe52d5-metrics-certs\") pod \"network-metrics-daemon-mqmtd\" (UID: \"be985ff1-de8a-431f-a230-67894ebe52d5\") " pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:10:22 crc kubenswrapper[4948]: I0312 00:10:22.359608 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-apiserver/apiserver-76f77b778f-6jfc4" Mar 12 00:10:22 crc kubenswrapper[4948]: I0312 00:10:22.365409 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-apiserver/apiserver-76f77b778f-6jfc4" Mar 12 00:10:22 crc kubenswrapper[4948]: E0312 00:10:22.436561 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 00:10:22.936537421 +0000 UTC m=+222.392141159 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g4kcb" (UID: "af75ce0a-d6bb-48db-a422-08b1d87e1d3f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:22 crc kubenswrapper[4948]: I0312 00:10:22.436176 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:22 crc kubenswrapper[4948]: I0312 00:10:22.439338 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:10:22 crc kubenswrapper[4948]: I0312 00:10:22.462038 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" Mar 12 00:10:22 crc kubenswrapper[4948]: I0312 00:10:22.468130 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" Mar 12 00:10:22 crc kubenswrapper[4948]: I0312 00:10:22.541893 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 00:10:22 crc kubenswrapper[4948]: E0312 00:10:22.542225 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 00:10:23.042210161 +0000 UTC m=+222.497813899 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:22 crc kubenswrapper[4948]: I0312 00:10:22.584763 4948 patch_prober.go:28] interesting pod/console-operator-58897d9998-z7w95 container/console-operator namespace/openshift-console-operator: Readiness probe status=failure output="Get \"https://10.217.0.35:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 12 00:10:22 crc kubenswrapper[4948]: I0312 00:10:22.584828 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console-operator/console-operator-58897d9998-z7w95" podUID="b265c22f-ab92-48e4-9c3e-db6ea6956f6f" containerName="console-operator" probeResult="failure" output="Get \"https://10.217.0.35:8443/readyz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 12 00:10:22 crc kubenswrapper[4948]: I0312 00:10:22.637960 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-multus/network-metrics-daemon-mqmtd" Mar 12 00:10:22 crc kubenswrapper[4948]: I0312 00:10:22.645999 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:22 crc kubenswrapper[4948]: E0312 00:10:22.646363 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 00:10:23.146350686 +0000 UTC m=+222.601954424 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g4kcb" (UID: "af75ce0a-d6bb-48db-a422-08b1d87e1d3f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:22 crc kubenswrapper[4948]: I0312 00:10:22.738533 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator-operator/kube-storage-version-migrator-operator-b67b599dd-kvh9w" event={"ID":"1666a214-546a-4063-a627-e990781797c8","Type":"ContainerStarted","Data":"326e37f4ea8b0b4a56d39330e17442b42113a01aa9ea18902b1a5f0ccd10bd80"} Mar 12 00:10:22 crc kubenswrapper[4948]: I0312 00:10:22.747052 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 00:10:22 crc kubenswrapper[4948]: E0312 00:10:22.748083 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 00:10:23.248067543 +0000 UTC m=+222.703671281 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:22 crc kubenswrapper[4948]: I0312 00:10:22.780960 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jlrln" event={"ID":"d33d5db8-961e-4c7a-950e-cb3f88a3f9b0","Type":"ContainerStarted","Data":"9bc07a72fc1315f3b054f9add4bd08307a4982221d8d0637cf58071a79e2f150"} Mar 12 00:10:22 crc kubenswrapper[4948]: I0312 00:10:22.781362 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jlrln" event={"ID":"d33d5db8-961e-4c7a-950e-cb3f88a3f9b0","Type":"ContainerStarted","Data":"75d0bfd379959fefdd0f8d2a2ee6b739eef63d6becf7a1196d7a8fcad3764a85"} Mar 12 00:10:22 crc kubenswrapper[4948]: I0312 00:10:22.781948 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jlrln" Mar 12 00:10:22 crc kubenswrapper[4948]: I0312 00:10:22.801833 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-z6v9h" event={"ID":"2b9e45c7-edbb-4cbb-9ee2-ae7efb61fbb3","Type":"ContainerStarted","Data":"b983b3e39cc45e095014877a6a4296cc9bcaee7a56fad0615b5b896aec6fec62"} Mar 12 00:10:22 crc kubenswrapper[4948]: I0312 00:10:22.815196 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jlrln" podStartSLOduration=183.815182602 podStartE2EDuration="3m3.815182602s" podCreationTimestamp="2026-03-12 00:07:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:10:22.813603565 +0000 UTC m=+222.269207303" watchObservedRunningTime="2026-03-12 00:10:22.815182602 +0000 UTC m=+222.270786340" Mar 12 00:10:22 crc kubenswrapper[4948]: I0312 00:10:22.847424 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-operator/ingress-operator-5b745b69d9-z6v9h" podStartSLOduration=183.847408426 podStartE2EDuration="3m3.847408426s" podCreationTimestamp="2026-03-12 00:07:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:10:22.845291886 +0000 UTC m=+222.300895624" watchObservedRunningTime="2026-03-12 00:10:22.847408426 +0000 UTC m=+222.303012154" Mar 12 00:10:22 crc kubenswrapper[4948]: I0312 00:10:22.848788 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:22 crc kubenswrapper[4948]: E0312 00:10:22.849892 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 00:10:23.349876413 +0000 UTC m=+222.805480151 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g4kcb" (UID: "af75ce0a-d6bb-48db-a422-08b1d87e1d3f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:22 crc kubenswrapper[4948]: I0312 00:10:22.868690 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-5q9rs" event={"ID":"4ba12474-c71f-429c-bcdf-9d74eb0b4b7a","Type":"ContainerStarted","Data":"3ee879c0e199c9f5537731ce079767a856df33be9330db385212470a2216947e"} Mar 12 00:10:22 crc kubenswrapper[4948]: I0312 00:10:22.883840 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rfxxc" event={"ID":"29cf53b1-0215-4dd4-a19f-24a48d4ded9d","Type":"ContainerStarted","Data":"3668990daae0096b7081f1f6afd5811f576c54cefefecf04455d65ccc401c38b"} Mar 12 00:10:22 crc kubenswrapper[4948]: I0312 00:10:22.883884 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rfxxc" event={"ID":"29cf53b1-0215-4dd4-a19f-24a48d4ded9d","Type":"ContainerStarted","Data":"34391a362e24396b5798b92401002fdc4dd02befc4391f0385d5cc6e3d3d763e"} Mar 12 00:10:22 crc kubenswrapper[4948]: I0312 00:10:22.914813 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-srr9m" event={"ID":"677936de-2686-46cf-99e9-cae2355c42b4","Type":"ContainerStarted","Data":"39b572e385797cb848a75684a308d5655c275fac74229d6bb162953b8e97f475"} Mar 12 00:10:22 crc kubenswrapper[4948]: I0312 00:10:22.914858 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-storage-version-migrator/migrator-59844c95c7-srr9m" event={"ID":"677936de-2686-46cf-99e9-cae2355c42b4","Type":"ContainerStarted","Data":"fdcb1ff4905e05b2ee8768f9720cc8151de974729a7d8d21a20d7a4a82fa9e0d"} Mar 12 00:10:22 crc kubenswrapper[4948]: I0312 00:10:22.944393 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-d24n2" event={"ID":"09e21136-b8c7-4680-ab71-73a86766a7b8","Type":"ContainerStarted","Data":"69bb2871e7ee21767e42be42f3ce1091d8f0858409091aafafbb5561358f8ffc"} Mar 12 00:10:22 crc kubenswrapper[4948]: I0312 00:10:22.944436 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-admission-controller-857f4d67dd-d24n2" event={"ID":"09e21136-b8c7-4680-ab71-73a86766a7b8","Type":"ContainerStarted","Data":"8052e1350ac1869e0b3105f46f4ed9794f083e1099f8169e5284649255c2d27b"} Mar 12 00:10:22 crc kubenswrapper[4948]: I0312 00:10:22.953656 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 00:10:22 crc kubenswrapper[4948]: E0312 00:10:22.954588 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 00:10:23.454573531 +0000 UTC m=+222.910177269 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:22 crc kubenswrapper[4948]: I0312 00:10:22.956173 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9k744" event={"ID":"1196fdb1-9c09-4982-9c37-35f46fabc1c5","Type":"ContainerStarted","Data":"c2c1b02a176662779c2956aad279066f95bf3f9485a09a2f7685ae34fe96afbb"} Mar 12 00:10:22 crc kubenswrapper[4948]: I0312 00:10:22.956206 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9k744" event={"ID":"1196fdb1-9c09-4982-9c37-35f46fabc1c5","Type":"ContainerStarted","Data":"a7fa54706ff89bce4249adbf991101eae8b99344f2cc04ea6eb71232b381e0f5"} Mar 12 00:10:22 crc kubenswrapper[4948]: I0312 00:10:22.965048 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h6z8v" event={"ID":"12c99e25-86bd-4923-ab44-55e571661c0f","Type":"ContainerStarted","Data":"d703559144efde320436014ef5fcd4bd499d9261cebfeac6da6561de5ab4f251"} Mar 12 00:10:22 crc kubenswrapper[4948]: I0312 00:10:22.966259 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h6z8v" Mar 12 00:10:22 crc kubenswrapper[4948]: I0312 00:10:22.970381 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-cluster-samples-operator/cluster-samples-operator-665b6dd947-rfxxc" podStartSLOduration=183.9703624 podStartE2EDuration="3m3.9703624s" podCreationTimestamp="2026-03-12 00:07:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:10:22.966591912 +0000 UTC m=+222.422195650" watchObservedRunningTime="2026-03-12 00:10:22.9703624 +0000 UTC m=+222.425966148" Mar 12 00:10:22 crc kubenswrapper[4948]: I0312 00:10:22.989180 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lkhdg" event={"ID":"bd97cad6-0d6d-4bec-9d2f-e17b03b220f6","Type":"ContainerStarted","Data":"fc670f19aee5b39097f0b308ca5cf5bd6e939197c2ee507b259e1b05df4af8ee"} Mar 12 00:10:22 crc kubenswrapper[4948]: I0312 00:10:22.989231 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lkhdg" event={"ID":"bd97cad6-0d6d-4bec-9d2f-e17b03b220f6","Type":"ContainerStarted","Data":"64655603cf1b4bfa8e2a7f01eaca9fd06335116969fe7ec563122262d41594f1"} Mar 12 00:10:23 crc kubenswrapper[4948]: I0312 00:10:23.001924 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/multus-admission-controller-857f4d67dd-d24n2" podStartSLOduration=184.001656762 podStartE2EDuration="3m4.001656762s" podCreationTimestamp="2026-03-12 00:07:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:10:22.996319257 +0000 UTC m=+222.451922995" watchObservedRunningTime="2026-03-12 00:10:23.001656762 +0000 UTC m=+222.457260500" Mar 12 00:10:23 crc kubenswrapper[4948]: I0312 00:10:23.012019 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/olm-operator-6b444d44fb-h6z8v" Mar 12 00:10:23 crc kubenswrapper[4948]: I0312 00:10:23.014010 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-v59sv" event={"ID":"7d1dc3d3-060c-4aba-aa54-0a3ea767f299","Type":"ContainerStarted","Data":"0c8782486aa66eec079c20339cd2453ac85aaf9442465e14bc81dd0ac12c7e67"} Mar 12 00:10:23 crc kubenswrapper[4948]: I0312 00:10:23.020440 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dc6wb" event={"ID":"7599787e-ea15-40ca-b365-fdf406e3c337","Type":"ContainerStarted","Data":"64950f33c098351edb300f07c3c824c17195265abe8fedb01c38d3541549c034"} Mar 12 00:10:23 crc kubenswrapper[4948]: I0312 00:10:23.021103 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dc6wb" Mar 12 00:10:23 crc kubenswrapper[4948]: I0312 00:10:23.055037 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:23 crc kubenswrapper[4948]: E0312 00:10:23.058853 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 00:10:23.558841978 +0000 UTC m=+223.014445706 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g4kcb" (UID: "af75ce0a-d6bb-48db-a422-08b1d87e1d3f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:23 crc kubenswrapper[4948]: I0312 00:10:23.063449 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29554560-8mgv9" event={"ID":"5b801839-e39a-48a9-9eee-1bb95df5fbfb","Type":"ContainerStarted","Data":"542707e4c866aa968cbd0962bee2140bff37118c8f12f28bc2e587b0cd01b2d5"} Mar 12 00:10:23 crc kubenswrapper[4948]: I0312 00:10:23.093376 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-operator-74547568cd-9k744" podStartSLOduration=184.093360316 podStartE2EDuration="3m4.093360316s" podCreationTimestamp="2026-03-12 00:07:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:10:23.053807821 +0000 UTC m=+222.509411599" watchObservedRunningTime="2026-03-12 00:10:23.093360316 +0000 UTC m=+222.548964054" Mar 12 00:10:23 crc kubenswrapper[4948]: I0312 00:10:23.093626 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6m4k7" event={"ID":"d08b0c90-3166-4919-88b3-345f5be3efe7","Type":"ContainerStarted","Data":"eb90bbad5a19b97f7810e350de4431d704ab96ed36497838fc5b5e34208c695d"} Mar 12 00:10:23 crc kubenswrapper[4948]: I0312 00:10:23.105920 4948 patch_prober.go:28] interesting pod/router-default-5444994796-7vhkr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 12 00:10:23 crc kubenswrapper[4948]: [-]has-synced failed: reason withheld Mar 12 00:10:23 crc kubenswrapper[4948]: [+]process-running ok Mar 12 00:10:23 crc kubenswrapper[4948]: healthz check failed Mar 12 00:10:23 crc kubenswrapper[4948]: I0312 00:10:23.105978 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7vhkr" podUID="e6184ca8-dc6e-4d34-ad0e-87c7107acb82" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 12 00:10:23 crc kubenswrapper[4948]: I0312 00:10:23.129108 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dc6wb" Mar 12 00:10:23 crc kubenswrapper[4948]: I0312 00:10:23.141308 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-apiserver/apiserver-76f77b778f-6jfc4" event={"ID":"2040ed8d-ae32-4909-9161-bc7ffbe49edd","Type":"ContainerStarted","Data":"9d0e3e91509f25f68e122b1f1b23acee4185d7096d340ee67b61b5efb8692b9b"} Mar 12 00:10:23 crc kubenswrapper[4948]: I0312 00:10:23.150615 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/catalog-operator-68c6474976-dc6wb" podStartSLOduration=184.150597634 podStartE2EDuration="3m4.150597634s" podCreationTimestamp="2026-03-12 00:07:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:10:23.139655308 +0000 UTC m=+222.595259046" watchObservedRunningTime="2026-03-12 00:10:23.150597634 +0000 UTC m=+222.606201372" Mar 12 00:10:23 crc kubenswrapper[4948]: I0312 00:10:23.163537 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 00:10:23 crc kubenswrapper[4948]: E0312 00:10:23.164536 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 00:10:23.664515649 +0000 UTC m=+223.120119377 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:23 crc kubenswrapper[4948]: I0312 00:10:23.185257 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ingress-canary/ingress-canary-mdhf9" event={"ID":"15828966-e769-43d4-b4a3-7a509fa72814","Type":"ContainerStarted","Data":"6f48ba23171e31f0628a2411fe68fe28eb5797457729670cfad796f80f9b54c0"} Mar 12 00:10:23 crc kubenswrapper[4948]: I0312 00:10:23.205867 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-controller-manager-operator/kube-controller-manager-operator-78b949d7b-v59sv" podStartSLOduration=184.205826014 podStartE2EDuration="3m4.205826014s" podCreationTimestamp="2026-03-12 00:07:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:10:23.204548434 +0000 UTC m=+222.660152172" watchObservedRunningTime="2026-03-12 00:10:23.205826014 +0000 UTC m=+222.661429752" Mar 12 00:10:23 crc kubenswrapper[4948]: I0312 00:10:23.206179 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-machine-config-operator/machine-config-controller-84d6567774-lkhdg" podStartSLOduration=184.206175673 podStartE2EDuration="3m4.206175673s" podCreationTimestamp="2026-03-12 00:07:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:10:23.173579431 +0000 UTC m=+222.629183159" watchObservedRunningTime="2026-03-12 00:10:23.206175673 +0000 UTC m=+222.661779411" Mar 12 00:10:23 crc kubenswrapper[4948]: I0312 00:10:23.211896 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-prqnt" event={"ID":"2851d901-f861-432b-a61e-5e0391f1b2b0","Type":"ContainerStarted","Data":"9308c18d5557505f9dc32ab60ec823355f3870aebc0d88c237556fb279559795"} Mar 12 00:10:23 crc kubenswrapper[4948]: I0312 00:10:23.211938 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-dns/dns-default-prqnt" event={"ID":"2851d901-f861-432b-a61e-5e0391f1b2b0","Type":"ContainerStarted","Data":"4b592ad9ecaf9e12c9bdb90451bbe3d1621b9101461da6da8c47747bc6881f34"} Mar 12 00:10:23 crc kubenswrapper[4948]: I0312 00:10:23.212549 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-dns/dns-default-prqnt" Mar 12 00:10:23 crc kubenswrapper[4948]: I0312 00:10:23.216068 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5qjdn" event={"ID":"3e538a31-98b5-4353-8038-4a7580be89dd","Type":"ContainerStarted","Data":"12d04130ce50bc2846ca67d8dac47c4043c21f2441d1ba4caac8b41f11b7206a"} Mar 12 00:10:23 crc kubenswrapper[4948]: I0312 00:10:23.217102 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5qjdn" Mar 12 00:10:23 crc kubenswrapper[4948]: I0312 00:10:23.218529 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver-operator/kube-apiserver-operator-766d6c64bb-gc7vl" event={"ID":"4dc52af7-648a-40e7-bdae-50672f5143dc","Type":"ContainerStarted","Data":"dd21a03459cab65d03b7346218abd065ebb5408d55d44293e77ec3133238a267"} Mar 12 00:10:23 crc kubenswrapper[4948]: I0312 00:10:23.220097 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-service-ca/service-ca-9c57cc56f-q79m4" event={"ID":"141bb112-1f07-4db5-9a4e-b7d94f12c5e2","Type":"ContainerStarted","Data":"88bc2927d9d52f191fc927cf42e9375f70b7c8cb9aefba552b7d6d8bb0996cce"} Mar 12 00:10:23 crc kubenswrapper[4948]: I0312 00:10:23.221473 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-879f6c89f-9tpt4" podUID="9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5" containerName="controller-manager" containerID="cri-o://6d8adce161c66b17114e112a0fc6f4519e1429afd2849dbda606aa4048b581b2" gracePeriod=30 Mar 12 00:10:23 crc kubenswrapper[4948]: I0312 00:10:23.224162 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rp8sv" podUID="60024b0b-f817-477f-83f1-e53cbc1e05b0" containerName="route-controller-manager" containerID="cri-o://567bd37219624454e44e80dc908b61d93b5a4feeeecc47a009a525503103bc90" gracePeriod=30 Mar 12 00:10:23 crc kubenswrapper[4948]: I0312 00:10:23.225139 4948 patch_prober.go:28] interesting pod/marketplace-operator-79b997595-zkbhv container/marketplace-operator namespace/openshift-marketplace: Readiness probe status=failure output="Get \"http://10.217.0.40:8080/healthz\": dial tcp 10.217.0.40:8080: connect: connection refused" start-of-body= Mar 12 00:10:23 crc kubenswrapper[4948]: I0312 00:10:23.230296 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-marketplace/marketplace-operator-79b997595-zkbhv" podUID="db40ba09-a72f-43d5-8037-e71c96e2bf5b" containerName="marketplace-operator" probeResult="failure" output="Get \"http://10.217.0.40:8080/healthz\": dial tcp 10.217.0.40:8080: connect: connection refused" Mar 12 00:10:23 crc kubenswrapper[4948]: I0312 00:10:23.256535 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29554560-8mgv9" podStartSLOduration=184.25652065 podStartE2EDuration="3m4.25652065s" podCreationTimestamp="2026-03-12 00:07:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:10:23.254763549 +0000 UTC m=+222.710367287" watchObservedRunningTime="2026-03-12 00:10:23.25652065 +0000 UTC m=+222.712124388" Mar 12 00:10:23 crc kubenswrapper[4948]: I0312 00:10:23.260482 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-oauth-apiserver/apiserver-7bbb656c7d-fx7lk" Mar 12 00:10:23 crc kubenswrapper[4948]: I0312 00:10:23.270246 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:23 crc kubenswrapper[4948]: E0312 00:10:23.279422 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 00:10:23.779410425 +0000 UTC m=+223.235014163 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g4kcb" (UID: "af75ce0a-d6bb-48db-a422-08b1d87e1d3f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:23 crc kubenswrapper[4948]: I0312 00:10:23.372967 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 00:10:23 crc kubenswrapper[4948]: E0312 00:10:23.373444 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 00:10:23.873429113 +0000 UTC m=+223.329032851 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:23 crc kubenswrapper[4948]: I0312 00:10:23.453774 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console-operator/console-operator-58897d9998-z7w95" Mar 12 00:10:23 crc kubenswrapper[4948]: I0312 00:10:23.453954 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ingress-canary/ingress-canary-mdhf9" podStartSLOduration=8.453939236 podStartE2EDuration="8.453939236s" podCreationTimestamp="2026-03-12 00:10:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:10:23.366730396 +0000 UTC m=+222.822334134" watchObservedRunningTime="2026-03-12 00:10:23.453939236 +0000 UTC m=+222.909542974" Mar 12 00:10:23 crc kubenswrapper[4948]: I0312 00:10:23.457015 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-scheduler-operator/openshift-kube-scheduler-operator-5fdd9b5758-6m4k7" podStartSLOduration=184.457004257 podStartE2EDuration="3m4.457004257s" podCreationTimestamp="2026-03-12 00:07:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:10:23.45288131 +0000 UTC m=+222.908485048" watchObservedRunningTime="2026-03-12 00:10:23.457004257 +0000 UTC m=+222.912607995" Mar 12 00:10:23 crc kubenswrapper[4948]: I0312 00:10:23.475957 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:23 crc kubenswrapper[4948]: E0312 00:10:23.476335 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 00:10:23.976307418 +0000 UTC m=+223.431911156 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g4kcb" (UID: "af75ce0a-d6bb-48db-a422-08b1d87e1d3f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:23 crc kubenswrapper[4948]: I0312 00:10:23.557739 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-service-ca/service-ca-9c57cc56f-q79m4" podStartSLOduration=184.557723732 podStartE2EDuration="3m4.557723732s" podCreationTimestamp="2026-03-12 00:07:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:10:23.496212774 +0000 UTC m=+222.951816512" watchObservedRunningTime="2026-03-12 00:10:23.557723732 +0000 UTC m=+223.013327470" Mar 12 00:10:23 crc kubenswrapper[4948]: I0312 00:10:23.559773 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-dns/dns-default-prqnt" podStartSLOduration=8.559769439 podStartE2EDuration="8.559769439s" podCreationTimestamp="2026-03-12 00:10:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:10:23.558617872 +0000 UTC m=+223.014221610" watchObservedRunningTime="2026-03-12 00:10:23.559769439 +0000 UTC m=+223.015373167" Mar 12 00:10:23 crc kubenswrapper[4948]: I0312 00:10:23.578901 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 00:10:23 crc kubenswrapper[4948]: E0312 00:10:23.579199 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 00:10:24.079183273 +0000 UTC m=+223.534787011 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:23 crc kubenswrapper[4948]: I0312 00:10:23.623888 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5qjdn" podStartSLOduration=184.623872948 podStartE2EDuration="3m4.623872948s" podCreationTimestamp="2026-03-12 00:07:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:10:23.621759558 +0000 UTC m=+223.077363296" watchObservedRunningTime="2026-03-12 00:10:23.623872948 +0000 UTC m=+223.079476706" Mar 12 00:10:23 crc kubenswrapper[4948]: I0312 00:10:23.654459 4948 patch_prober.go:28] interesting pod/apiserver-76f77b778f-6jfc4 container/openshift-apiserver namespace/openshift-apiserver: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[+]ping ok Mar 12 00:10:23 crc kubenswrapper[4948]: [+]log ok Mar 12 00:10:23 crc kubenswrapper[4948]: [+]etcd ok Mar 12 00:10:23 crc kubenswrapper[4948]: [+]poststarthook/start-apiserver-admission-initializer ok Mar 12 00:10:23 crc kubenswrapper[4948]: [+]poststarthook/generic-apiserver-start-informers ok Mar 12 00:10:23 crc kubenswrapper[4948]: [+]poststarthook/max-in-flight-filter ok Mar 12 00:10:23 crc kubenswrapper[4948]: [+]poststarthook/storage-object-count-tracker-hook ok Mar 12 00:10:23 crc kubenswrapper[4948]: [+]poststarthook/image.openshift.io-apiserver-caches ok Mar 12 00:10:23 crc kubenswrapper[4948]: [-]poststarthook/authorization.openshift.io-bootstrapclusterroles failed: reason withheld Mar 12 00:10:23 crc kubenswrapper[4948]: [-]poststarthook/authorization.openshift.io-ensurenodebootstrap-sa failed: reason withheld Mar 12 00:10:23 crc kubenswrapper[4948]: [+]poststarthook/project.openshift.io-projectcache ok Mar 12 00:10:23 crc kubenswrapper[4948]: [+]poststarthook/project.openshift.io-projectauthorizationcache ok Mar 12 00:10:23 crc kubenswrapper[4948]: [+]poststarthook/openshift.io-startinformers ok Mar 12 00:10:23 crc kubenswrapper[4948]: [+]poststarthook/openshift.io-restmapperupdater ok Mar 12 00:10:23 crc kubenswrapper[4948]: [+]poststarthook/quota.openshift.io-clusterquotamapping ok Mar 12 00:10:23 crc kubenswrapper[4948]: livez check failed Mar 12 00:10:23 crc kubenswrapper[4948]: I0312 00:10:23.654506 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-apiserver/apiserver-76f77b778f-6jfc4" podUID="2040ed8d-ae32-4909-9161-bc7ffbe49edd" containerName="openshift-apiserver" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 12 00:10:23 crc kubenswrapper[4948]: I0312 00:10:23.680210 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:23 crc kubenswrapper[4948]: E0312 00:10:23.680532 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 00:10:24.180520502 +0000 UTC m=+223.636124240 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g4kcb" (UID: "af75ce0a-d6bb-48db-a422-08b1d87e1d3f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:23 crc kubenswrapper[4948]: I0312 00:10:23.697219 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-multus/network-metrics-daemon-mqmtd"] Mar 12 00:10:23 crc kubenswrapper[4948]: I0312 00:10:23.780195 4948 ???:1] "http: TLS handshake error from 192.168.126.11:35550: no serving certificate available for the kubelet" Mar 12 00:10:23 crc kubenswrapper[4948]: I0312 00:10:23.781871 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 00:10:23 crc kubenswrapper[4948]: E0312 00:10:23.782274 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 00:10:24.28226028 +0000 UTC m=+223.737864018 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:23 crc kubenswrapper[4948]: W0312 00:10:23.875446 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbe985ff1_de8a_431f_a230_67894ebe52d5.slice/crio-10411dbe8d718cbfaa967631346d9f78802e245fc4667a722de79cb6482e6b73 WatchSource:0}: Error finding container 10411dbe8d718cbfaa967631346d9f78802e245fc4667a722de79cb6482e6b73: Status 404 returned error can't find the container with id 10411dbe8d718cbfaa967631346d9f78802e245fc4667a722de79cb6482e6b73 Mar 12 00:10:23 crc kubenswrapper[4948]: I0312 00:10:23.886202 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:23 crc kubenswrapper[4948]: E0312 00:10:23.886550 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 00:10:24.386536219 +0000 UTC m=+223.842139957 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g4kcb" (UID: "af75ce0a-d6bb-48db-a422-08b1d87e1d3f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:23 crc kubenswrapper[4948]: I0312 00:10:23.981392 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-9tpt4" Mar 12 00:10:23 crc kubenswrapper[4948]: I0312 00:10:23.990247 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 00:10:23 crc kubenswrapper[4948]: E0312 00:10:23.990615 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 00:10:24.490600291 +0000 UTC m=+223.946204029 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.013727 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7bb5cdccc6-pwdcd"] Mar 12 00:10:24 crc kubenswrapper[4948]: E0312 00:10:24.013921 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5" containerName="controller-manager" Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.013934 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5" containerName="controller-manager" Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.014071 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5" containerName="controller-manager" Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.014453 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7bb5cdccc6-pwdcd" Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.025995 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7bb5cdccc6-pwdcd"] Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.092691 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-2j6ml\" (UniqueName: \"kubernetes.io/projected/9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5-kube-api-access-2j6ml\") pod \"9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5\" (UID: \"9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5\") " Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.092741 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5-proxy-ca-bundles\") pod \"9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5\" (UID: \"9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5\") " Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.092778 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5-serving-cert\") pod \"9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5\" (UID: \"9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5\") " Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.092837 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5-config\") pod \"9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5\" (UID: \"9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5\") " Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.092974 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5-client-ca\") pod \"9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5\" (UID: \"9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5\") " Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.093099 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd346d37-ec14-49f0-8e88-f73930b8bab8-config\") pod \"controller-manager-7bb5cdccc6-pwdcd\" (UID: \"cd346d37-ec14-49f0-8e88-f73930b8bab8\") " pod="openshift-controller-manager/controller-manager-7bb5cdccc6-pwdcd" Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.093129 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.093152 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/cd346d37-ec14-49f0-8e88-f73930b8bab8-proxy-ca-bundles\") pod \"controller-manager-7bb5cdccc6-pwdcd\" (UID: \"cd346d37-ec14-49f0-8e88-f73930b8bab8\") " pod="openshift-controller-manager/controller-manager-7bb5cdccc6-pwdcd" Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.093176 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cd346d37-ec14-49f0-8e88-f73930b8bab8-serving-cert\") pod \"controller-manager-7bb5cdccc6-pwdcd\" (UID: \"cd346d37-ec14-49f0-8e88-f73930b8bab8\") " pod="openshift-controller-manager/controller-manager-7bb5cdccc6-pwdcd" Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.093207 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mplgq\" (UniqueName: \"kubernetes.io/projected/cd346d37-ec14-49f0-8e88-f73930b8bab8-kube-api-access-mplgq\") pod \"controller-manager-7bb5cdccc6-pwdcd\" (UID: \"cd346d37-ec14-49f0-8e88-f73930b8bab8\") " pod="openshift-controller-manager/controller-manager-7bb5cdccc6-pwdcd" Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.093236 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cd346d37-ec14-49f0-8e88-f73930b8bab8-client-ca\") pod \"controller-manager-7bb5cdccc6-pwdcd\" (UID: \"cd346d37-ec14-49f0-8e88-f73930b8bab8\") " pod="openshift-controller-manager/controller-manager-7bb5cdccc6-pwdcd" Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.094661 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5" (UID: "9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:10:24 crc kubenswrapper[4948]: E0312 00:10:24.097709 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 00:10:24.597691474 +0000 UTC m=+224.053295282 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g4kcb" (UID: "af75ce0a-d6bb-48db-a422-08b1d87e1d3f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.098449 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5-config" (OuterVolumeSpecName: "config") pod "9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5" (UID: "9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.098900 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5-client-ca" (OuterVolumeSpecName: "client-ca") pod "9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5" (UID: "9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.101515 4948 patch_prober.go:28] interesting pod/router-default-5444994796-7vhkr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 12 00:10:24 crc kubenswrapper[4948]: [-]has-synced failed: reason withheld Mar 12 00:10:24 crc kubenswrapper[4948]: [+]process-running ok Mar 12 00:10:24 crc kubenswrapper[4948]: healthz check failed Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.101549 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7vhkr" podUID="e6184ca8-dc6e-4d34-ad0e-87c7107acb82" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.107222 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5-kube-api-access-2j6ml" (OuterVolumeSpecName: "kube-api-access-2j6ml") pod "9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5" (UID: "9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5"). InnerVolumeSpecName "kube-api-access-2j6ml". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.108904 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5" (UID: "9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.195089 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.195249 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd346d37-ec14-49f0-8e88-f73930b8bab8-config\") pod \"controller-manager-7bb5cdccc6-pwdcd\" (UID: \"cd346d37-ec14-49f0-8e88-f73930b8bab8\") " pod="openshift-controller-manager/controller-manager-7bb5cdccc6-pwdcd" Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.195284 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/cd346d37-ec14-49f0-8e88-f73930b8bab8-proxy-ca-bundles\") pod \"controller-manager-7bb5cdccc6-pwdcd\" (UID: \"cd346d37-ec14-49f0-8e88-f73930b8bab8\") " pod="openshift-controller-manager/controller-manager-7bb5cdccc6-pwdcd" Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.195302 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cd346d37-ec14-49f0-8e88-f73930b8bab8-serving-cert\") pod \"controller-manager-7bb5cdccc6-pwdcd\" (UID: \"cd346d37-ec14-49f0-8e88-f73930b8bab8\") " pod="openshift-controller-manager/controller-manager-7bb5cdccc6-pwdcd" Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.195347 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mplgq\" (UniqueName: \"kubernetes.io/projected/cd346d37-ec14-49f0-8e88-f73930b8bab8-kube-api-access-mplgq\") pod \"controller-manager-7bb5cdccc6-pwdcd\" (UID: \"cd346d37-ec14-49f0-8e88-f73930b8bab8\") " pod="openshift-controller-manager/controller-manager-7bb5cdccc6-pwdcd" Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.195371 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cd346d37-ec14-49f0-8e88-f73930b8bab8-client-ca\") pod \"controller-manager-7bb5cdccc6-pwdcd\" (UID: \"cd346d37-ec14-49f0-8e88-f73930b8bab8\") " pod="openshift-controller-manager/controller-manager-7bb5cdccc6-pwdcd" Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.195416 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5-config\") on node \"crc\" DevicePath \"\"" Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.195429 4948 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5-client-ca\") on node \"crc\" DevicePath \"\"" Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.195439 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-2j6ml\" (UniqueName: \"kubernetes.io/projected/9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5-kube-api-access-2j6ml\") on node \"crc\" DevicePath \"\"" Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.195447 4948 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.195455 4948 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.196139 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cd346d37-ec14-49f0-8e88-f73930b8bab8-client-ca\") pod \"controller-manager-7bb5cdccc6-pwdcd\" (UID: \"cd346d37-ec14-49f0-8e88-f73930b8bab8\") " pod="openshift-controller-manager/controller-manager-7bb5cdccc6-pwdcd" Mar 12 00:10:24 crc kubenswrapper[4948]: E0312 00:10:24.197070 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 00:10:24.696987076 +0000 UTC m=+224.152590814 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.198324 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/cd346d37-ec14-49f0-8e88-f73930b8bab8-proxy-ca-bundles\") pod \"controller-manager-7bb5cdccc6-pwdcd\" (UID: \"cd346d37-ec14-49f0-8e88-f73930b8bab8\") " pod="openshift-controller-manager/controller-manager-7bb5cdccc6-pwdcd" Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.200566 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd346d37-ec14-49f0-8e88-f73930b8bab8-config\") pod \"controller-manager-7bb5cdccc6-pwdcd\" (UID: \"cd346d37-ec14-49f0-8e88-f73930b8bab8\") " pod="openshift-controller-manager/controller-manager-7bb5cdccc6-pwdcd" Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.202151 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cd346d37-ec14-49f0-8e88-f73930b8bab8-serving-cert\") pod \"controller-manager-7bb5cdccc6-pwdcd\" (UID: \"cd346d37-ec14-49f0-8e88-f73930b8bab8\") " pod="openshift-controller-manager/controller-manager-7bb5cdccc6-pwdcd" Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.214236 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mplgq\" (UniqueName: \"kubernetes.io/projected/cd346d37-ec14-49f0-8e88-f73930b8bab8-kube-api-access-mplgq\") pod \"controller-manager-7bb5cdccc6-pwdcd\" (UID: \"cd346d37-ec14-49f0-8e88-f73930b8bab8\") " pod="openshift-controller-manager/controller-manager-7bb5cdccc6-pwdcd" Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.217531 4948 patch_prober.go:28] interesting pod/packageserver-d55dfcdfc-5qjdn container/packageserver namespace/openshift-operator-lifecycle-manager: Readiness probe status=failure output="Get \"https://10.217.0.27:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.217603 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5qjdn" podUID="3e538a31-98b5-4353-8038-4a7580be89dd" containerName="packageserver" probeResult="failure" output="Get \"https://10.217.0.27:5443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.264927 4948 generic.go:334] "Generic (PLEG): container finished" podID="60024b0b-f817-477f-83f1-e53cbc1e05b0" containerID="567bd37219624454e44e80dc908b61d93b5a4feeeecc47a009a525503103bc90" exitCode=0 Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.264995 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rp8sv" event={"ID":"60024b0b-f817-477f-83f1-e53cbc1e05b0","Type":"ContainerDied","Data":"567bd37219624454e44e80dc908b61d93b5a4feeeecc47a009a525503103bc90"} Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.286925 4948 generic.go:334] "Generic (PLEG): container finished" podID="9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5" containerID="6d8adce161c66b17114e112a0fc6f4519e1429afd2849dbda606aa4048b581b2" exitCode=0 Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.286983 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-9tpt4" event={"ID":"9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5","Type":"ContainerDied","Data":"6d8adce161c66b17114e112a0fc6f4519e1429afd2849dbda606aa4048b581b2"} Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.287008 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-879f6c89f-9tpt4" event={"ID":"9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5","Type":"ContainerDied","Data":"500e98a6770d1487c1d9daeeb538a9a1984dab56b1e205303c446292ff336d69"} Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.287023 4948 scope.go:117] "RemoveContainer" containerID="6d8adce161c66b17114e112a0fc6f4519e1429afd2849dbda606aa4048b581b2" Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.287115 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-879f6c89f-9tpt4" Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.299454 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:24 crc kubenswrapper[4948]: E0312 00:10:24.299731 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 00:10:24.799718888 +0000 UTC m=+224.255322626 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g4kcb" (UID: "af75ce0a-d6bb-48db-a422-08b1d87e1d3f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.309006 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-mqmtd" event={"ID":"be985ff1-de8a-431f-a230-67894ebe52d5","Type":"ContainerStarted","Data":"10411dbe8d718cbfaa967631346d9f78802e245fc4667a722de79cb6482e6b73"} Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.310413 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"5016c4612964f46ff7b4eb2ffc4867abcc1ebb2ce48b47a63579caad88422258"} Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.313858 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"71d53b60c9e7d4acd758a3be6daaa408b6d4aeb83fb746dd9533699614a8ea8c"} Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.313890 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-target-xd92c" event={"ID":"3b6479f0-333b-4a96-9adf-2099afdc2447","Type":"ContainerStarted","Data":"2330eee6f432410e9a1f9f0cf2269a3f363be4697a7a31648b1ae069efa5c373"} Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.314417 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.315712 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"d7fa884a65f992deba4fb7285cd73e876048a0cf2f420b744d686aacdb5ca8de"} Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.315729 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-diagnostics/network-check-source-55646444c4-trplf" event={"ID":"9d751cbb-f2e2-430d-9754-c882a5e924a5","Type":"ContainerStarted","Data":"a452549c6f21e3f4c0178726e719854e71db1aab1c1cc67944b4618d2dbaa217"} Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.332150 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-zkbhv" Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.338828 4948 scope.go:117] "RemoveContainer" containerID="6d8adce161c66b17114e112a0fc6f4519e1429afd2849dbda606aa4048b581b2" Mar 12 00:10:24 crc kubenswrapper[4948]: E0312 00:10:24.339333 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6d8adce161c66b17114e112a0fc6f4519e1429afd2849dbda606aa4048b581b2\": container with ID starting with 6d8adce161c66b17114e112a0fc6f4519e1429afd2849dbda606aa4048b581b2 not found: ID does not exist" containerID="6d8adce161c66b17114e112a0fc6f4519e1429afd2849dbda606aa4048b581b2" Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.339373 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6d8adce161c66b17114e112a0fc6f4519e1429afd2849dbda606aa4048b581b2"} err="failed to get container status \"6d8adce161c66b17114e112a0fc6f4519e1429afd2849dbda606aa4048b581b2\": rpc error: code = NotFound desc = could not find container \"6d8adce161c66b17114e112a0fc6f4519e1429afd2849dbda606aa4048b581b2\": container with ID starting with 6d8adce161c66b17114e112a0fc6f4519e1429afd2849dbda606aa4048b581b2 not found: ID does not exist" Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.345526 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rp8sv" Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.356747 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7bb5cdccc6-pwdcd" Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.357514 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/packageserver-d55dfcdfc-5qjdn" Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.395227 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-9tpt4"] Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.395961 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-879f6c89f-9tpt4"] Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.408917 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.409013 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60024b0b-f817-477f-83f1-e53cbc1e05b0-config\") pod \"60024b0b-f817-477f-83f1-e53cbc1e05b0\" (UID: \"60024b0b-f817-477f-83f1-e53cbc1e05b0\") " Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.409048 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6q7g6\" (UniqueName: \"kubernetes.io/projected/60024b0b-f817-477f-83f1-e53cbc1e05b0-kube-api-access-6q7g6\") pod \"60024b0b-f817-477f-83f1-e53cbc1e05b0\" (UID: \"60024b0b-f817-477f-83f1-e53cbc1e05b0\") " Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.409072 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/60024b0b-f817-477f-83f1-e53cbc1e05b0-serving-cert\") pod \"60024b0b-f817-477f-83f1-e53cbc1e05b0\" (UID: \"60024b0b-f817-477f-83f1-e53cbc1e05b0\") " Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.409112 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/60024b0b-f817-477f-83f1-e53cbc1e05b0-client-ca\") pod \"60024b0b-f817-477f-83f1-e53cbc1e05b0\" (UID: \"60024b0b-f817-477f-83f1-e53cbc1e05b0\") " Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.410712 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60024b0b-f817-477f-83f1-e53cbc1e05b0-client-ca" (OuterVolumeSpecName: "client-ca") pod "60024b0b-f817-477f-83f1-e53cbc1e05b0" (UID: "60024b0b-f817-477f-83f1-e53cbc1e05b0"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:10:24 crc kubenswrapper[4948]: E0312 00:10:24.411052 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 00:10:24.910992799 +0000 UTC m=+224.366596537 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.411583 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/60024b0b-f817-477f-83f1-e53cbc1e05b0-config" (OuterVolumeSpecName: "config") pod "60024b0b-f817-477f-83f1-e53cbc1e05b0" (UID: "60024b0b-f817-477f-83f1-e53cbc1e05b0"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.447301 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/60024b0b-f817-477f-83f1-e53cbc1e05b0-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "60024b0b-f817-477f-83f1-e53cbc1e05b0" (UID: "60024b0b-f817-477f-83f1-e53cbc1e05b0"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.447710 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/60024b0b-f817-477f-83f1-e53cbc1e05b0-kube-api-access-6q7g6" (OuterVolumeSpecName: "kube-api-access-6q7g6") pod "60024b0b-f817-477f-83f1-e53cbc1e05b0" (UID: "60024b0b-f817-477f-83f1-e53cbc1e05b0"). InnerVolumeSpecName "kube-api-access-6q7g6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.512677 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.512775 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/60024b0b-f817-477f-83f1-e53cbc1e05b0-config\") on node \"crc\" DevicePath \"\"" Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.512790 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6q7g6\" (UniqueName: \"kubernetes.io/projected/60024b0b-f817-477f-83f1-e53cbc1e05b0-kube-api-access-6q7g6\") on node \"crc\" DevicePath \"\"" Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.512828 4948 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/60024b0b-f817-477f-83f1-e53cbc1e05b0-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.512841 4948 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/60024b0b-f817-477f-83f1-e53cbc1e05b0-client-ca\") on node \"crc\" DevicePath \"\"" Mar 12 00:10:24 crc kubenswrapper[4948]: E0312 00:10:24.514332 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 00:10:25.014279903 +0000 UTC m=+224.469883641 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g4kcb" (UID: "af75ce0a-d6bb-48db-a422-08b1d87e1d3f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.613908 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 00:10:24 crc kubenswrapper[4948]: E0312 00:10:24.614109 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 00:10:25.114078676 +0000 UTC m=+224.569682424 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.614177 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:24 crc kubenswrapper[4948]: E0312 00:10:24.614483 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 00:10:25.114471015 +0000 UTC m=+224.570074753 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g4kcb" (UID: "af75ce0a-d6bb-48db-a422-08b1d87e1d3f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.638439 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-v8h88"] Mar 12 00:10:24 crc kubenswrapper[4948]: E0312 00:10:24.638955 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="60024b0b-f817-477f-83f1-e53cbc1e05b0" containerName="route-controller-manager" Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.638966 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="60024b0b-f817-477f-83f1-e53cbc1e05b0" containerName="route-controller-manager" Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.639081 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="60024b0b-f817-477f-83f1-e53cbc1e05b0" containerName="route-controller-manager" Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.640028 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v8h88" Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.646162 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.650520 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-v8h88"] Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.714998 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 00:10:24 crc kubenswrapper[4948]: E0312 00:10:24.715160 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 00:10:25.215132948 +0000 UTC m=+224.670736686 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.715221 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lbkcx\" (UniqueName: \"kubernetes.io/projected/5d83d568-3968-4ea6-b548-511c951defa5-kube-api-access-lbkcx\") pod \"community-operators-v8h88\" (UID: \"5d83d568-3968-4ea6-b548-511c951defa5\") " pod="openshift-marketplace/community-operators-v8h88" Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.715335 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5d83d568-3968-4ea6-b548-511c951defa5-utilities\") pod \"community-operators-v8h88\" (UID: \"5d83d568-3968-4ea6-b548-511c951defa5\") " pod="openshift-marketplace/community-operators-v8h88" Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.715367 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5d83d568-3968-4ea6-b548-511c951defa5-catalog-content\") pod \"community-operators-v8h88\" (UID: \"5d83d568-3968-4ea6-b548-511c951defa5\") " pod="openshift-marketplace/community-operators-v8h88" Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.715412 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:24 crc kubenswrapper[4948]: E0312 00:10:24.715636 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 00:10:25.21562269 +0000 UTC m=+224.671226428 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g4kcb" (UID: "af75ce0a-d6bb-48db-a422-08b1d87e1d3f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.718409 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7bb5cdccc6-pwdcd"] Mar 12 00:10:24 crc kubenswrapper[4948]: W0312 00:10:24.753273 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcd346d37_ec14_49f0_8e88_f73930b8bab8.slice/crio-4ef442baa04f491a58e34e7e726e281ce5c5a6d127017aca94eaf63dceefc1f0 WatchSource:0}: Error finding container 4ef442baa04f491a58e34e7e726e281ce5c5a6d127017aca94eaf63dceefc1f0: Status 404 returned error can't find the container with id 4ef442baa04f491a58e34e7e726e281ce5c5a6d127017aca94eaf63dceefc1f0 Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.816584 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.817143 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lbkcx\" (UniqueName: \"kubernetes.io/projected/5d83d568-3968-4ea6-b548-511c951defa5-kube-api-access-lbkcx\") pod \"community-operators-v8h88\" (UID: \"5d83d568-3968-4ea6-b548-511c951defa5\") " pod="openshift-marketplace/community-operators-v8h88" Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.817652 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5d83d568-3968-4ea6-b548-511c951defa5-utilities\") pod \"community-operators-v8h88\" (UID: \"5d83d568-3968-4ea6-b548-511c951defa5\") " pod="openshift-marketplace/community-operators-v8h88" Mar 12 00:10:24 crc kubenswrapper[4948]: E0312 00:10:24.817855 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 00:10:25.31783113 +0000 UTC m=+224.773434868 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.818205 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5d83d568-3968-4ea6-b548-511c951defa5-catalog-content\") pod \"community-operators-v8h88\" (UID: \"5d83d568-3968-4ea6-b548-511c951defa5\") " pod="openshift-marketplace/community-operators-v8h88" Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.818585 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5d83d568-3968-4ea6-b548-511c951defa5-utilities\") pod \"community-operators-v8h88\" (UID: \"5d83d568-3968-4ea6-b548-511c951defa5\") " pod="openshift-marketplace/community-operators-v8h88" Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.820595 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5d83d568-3968-4ea6-b548-511c951defa5-catalog-content\") pod \"community-operators-v8h88\" (UID: \"5d83d568-3968-4ea6-b548-511c951defa5\") " pod="openshift-marketplace/community-operators-v8h88" Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.828953 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-rflhd"] Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.829869 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rflhd" Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.832371 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.844381 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rflhd"] Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.848557 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lbkcx\" (UniqueName: \"kubernetes.io/projected/5d83d568-3968-4ea6-b548-511c951defa5-kube-api-access-lbkcx\") pod \"community-operators-v8h88\" (UID: \"5d83d568-3968-4ea6-b548-511c951defa5\") " pod="openshift-marketplace/community-operators-v8h88" Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.922809 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rljd9\" (UniqueName: \"kubernetes.io/projected/5184125a-dde1-4db4-9d6b-5b3c45d8d82a-kube-api-access-rljd9\") pod \"certified-operators-rflhd\" (UID: \"5184125a-dde1-4db4-9d6b-5b3c45d8d82a\") " pod="openshift-marketplace/certified-operators-rflhd" Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.922889 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.922916 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5184125a-dde1-4db4-9d6b-5b3c45d8d82a-catalog-content\") pod \"certified-operators-rflhd\" (UID: \"5184125a-dde1-4db4-9d6b-5b3c45d8d82a\") " pod="openshift-marketplace/certified-operators-rflhd" Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.922988 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5184125a-dde1-4db4-9d6b-5b3c45d8d82a-utilities\") pod \"certified-operators-rflhd\" (UID: \"5184125a-dde1-4db4-9d6b-5b3c45d8d82a\") " pod="openshift-marketplace/certified-operators-rflhd" Mar 12 00:10:24 crc kubenswrapper[4948]: E0312 00:10:24.925037 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 00:10:25.425023985 +0000 UTC m=+224.880627723 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g4kcb" (UID: "af75ce0a-d6bb-48db-a422-08b1d87e1d3f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.934509 4948 plugin_watcher.go:194] "Adding socket path or updating timestamp to desired state cache" path="/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock" Mar 12 00:10:24 crc kubenswrapper[4948]: I0312 00:10:24.960513 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v8h88" Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.025734 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-62wnf"] Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.026180 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.026328 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rljd9\" (UniqueName: \"kubernetes.io/projected/5184125a-dde1-4db4-9d6b-5b3c45d8d82a-kube-api-access-rljd9\") pod \"certified-operators-rflhd\" (UID: \"5184125a-dde1-4db4-9d6b-5b3c45d8d82a\") " pod="openshift-marketplace/certified-operators-rflhd" Mar 12 00:10:25 crc kubenswrapper[4948]: E0312 00:10:25.026377 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 00:10:25.526351735 +0000 UTC m=+224.981955473 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.026435 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.026490 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5184125a-dde1-4db4-9d6b-5b3c45d8d82a-catalog-content\") pod \"certified-operators-rflhd\" (UID: \"5184125a-dde1-4db4-9d6b-5b3c45d8d82a\") " pod="openshift-marketplace/certified-operators-rflhd" Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.026568 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5184125a-dde1-4db4-9d6b-5b3c45d8d82a-utilities\") pod \"certified-operators-rflhd\" (UID: \"5184125a-dde1-4db4-9d6b-5b3c45d8d82a\") " pod="openshift-marketplace/certified-operators-rflhd" Mar 12 00:10:25 crc kubenswrapper[4948]: E0312 00:10:25.026731 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 00:10:25.526716733 +0000 UTC m=+224.982320471 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g4kcb" (UID: "af75ce0a-d6bb-48db-a422-08b1d87e1d3f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.026970 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5184125a-dde1-4db4-9d6b-5b3c45d8d82a-catalog-content\") pod \"certified-operators-rflhd\" (UID: \"5184125a-dde1-4db4-9d6b-5b3c45d8d82a\") " pod="openshift-marketplace/certified-operators-rflhd" Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.027068 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5184125a-dde1-4db4-9d6b-5b3c45d8d82a-utilities\") pod \"certified-operators-rflhd\" (UID: \"5184125a-dde1-4db4-9d6b-5b3c45d8d82a\") " pod="openshift-marketplace/certified-operators-rflhd" Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.027583 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-62wnf" Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.040568 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-62wnf"] Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.043045 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rljd9\" (UniqueName: \"kubernetes.io/projected/5184125a-dde1-4db4-9d6b-5b3c45d8d82a-kube-api-access-rljd9\") pod \"certified-operators-rflhd\" (UID: \"5184125a-dde1-4db4-9d6b-5b3c45d8d82a\") " pod="openshift-marketplace/certified-operators-rflhd" Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.097847 4948 patch_prober.go:28] interesting pod/router-default-5444994796-7vhkr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 12 00:10:25 crc kubenswrapper[4948]: [-]has-synced failed: reason withheld Mar 12 00:10:25 crc kubenswrapper[4948]: [+]process-running ok Mar 12 00:10:25 crc kubenswrapper[4948]: healthz check failed Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.097929 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7vhkr" podUID="e6184ca8-dc6e-4d34-ad0e-87c7107acb82" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.127420 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 00:10:25 crc kubenswrapper[4948]: E0312 00:10:25.127679 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 00:10:25.627633652 +0000 UTC m=+225.083237400 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.128000 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.128032 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6318b61-ef2d-41da-8c92-ff7523ce613f-utilities\") pod \"community-operators-62wnf\" (UID: \"e6318b61-ef2d-41da-8c92-ff7523ce613f\") " pod="openshift-marketplace/community-operators-62wnf" Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.128057 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6318b61-ef2d-41da-8c92-ff7523ce613f-catalog-content\") pod \"community-operators-62wnf\" (UID: \"e6318b61-ef2d-41da-8c92-ff7523ce613f\") " pod="openshift-marketplace/community-operators-62wnf" Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.128101 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5gmsv\" (UniqueName: \"kubernetes.io/projected/e6318b61-ef2d-41da-8c92-ff7523ce613f-kube-api-access-5gmsv\") pod \"community-operators-62wnf\" (UID: \"e6318b61-ef2d-41da-8c92-ff7523ce613f\") " pod="openshift-marketplace/community-operators-62wnf" Mar 12 00:10:25 crc kubenswrapper[4948]: E0312 00:10:25.128367 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 00:10:25.628354658 +0000 UTC m=+225.083958396 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g4kcb" (UID: "af75ce0a-d6bb-48db-a422-08b1d87e1d3f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.162895 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rflhd" Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.230023 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.230270 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6318b61-ef2d-41da-8c92-ff7523ce613f-utilities\") pod \"community-operators-62wnf\" (UID: \"e6318b61-ef2d-41da-8c92-ff7523ce613f\") " pod="openshift-marketplace/community-operators-62wnf" Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.230295 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6318b61-ef2d-41da-8c92-ff7523ce613f-catalog-content\") pod \"community-operators-62wnf\" (UID: \"e6318b61-ef2d-41da-8c92-ff7523ce613f\") " pod="openshift-marketplace/community-operators-62wnf" Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.230356 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5gmsv\" (UniqueName: \"kubernetes.io/projected/e6318b61-ef2d-41da-8c92-ff7523ce613f-kube-api-access-5gmsv\") pod \"community-operators-62wnf\" (UID: \"e6318b61-ef2d-41da-8c92-ff7523ce613f\") " pod="openshift-marketplace/community-operators-62wnf" Mar 12 00:10:25 crc kubenswrapper[4948]: E0312 00:10:25.230670 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 00:10:25.73065684 +0000 UTC m=+225.186260578 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.231501 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6318b61-ef2d-41da-8c92-ff7523ce613f-utilities\") pod \"community-operators-62wnf\" (UID: \"e6318b61-ef2d-41da-8c92-ff7523ce613f\") " pod="openshift-marketplace/community-operators-62wnf" Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.233569 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6318b61-ef2d-41da-8c92-ff7523ce613f-catalog-content\") pod \"community-operators-62wnf\" (UID: \"e6318b61-ef2d-41da-8c92-ff7523ce613f\") " pod="openshift-marketplace/community-operators-62wnf" Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.234599 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-hcvm5"] Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.235433 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hcvm5" Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.259342 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5gmsv\" (UniqueName: \"kubernetes.io/projected/e6318b61-ef2d-41da-8c92-ff7523ce613f-kube-api-access-5gmsv\") pod \"community-operators-62wnf\" (UID: \"e6318b61-ef2d-41da-8c92-ff7523ce613f\") " pod="openshift-marketplace/community-operators-62wnf" Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.275497 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-v8h88"] Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.320928 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5" path="/var/lib/kubelet/pods/9430c046-c6a0-4fb5-8a6e-c69f3eeaeda5/volumes" Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.321596 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hcvm5"] Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.332048 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/239cfe46-3a3a-4624-9aa2-ea7112f28947-utilities\") pod \"certified-operators-hcvm5\" (UID: \"239cfe46-3a3a-4624-9aa2-ea7112f28947\") " pod="openshift-marketplace/certified-operators-hcvm5" Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.332096 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/239cfe46-3a3a-4624-9aa2-ea7112f28947-catalog-content\") pod \"certified-operators-hcvm5\" (UID: \"239cfe46-3a3a-4624-9aa2-ea7112f28947\") " pod="openshift-marketplace/certified-operators-hcvm5" Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.332155 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-pk6rv\" (UniqueName: \"kubernetes.io/projected/239cfe46-3a3a-4624-9aa2-ea7112f28947-kube-api-access-pk6rv\") pod \"certified-operators-hcvm5\" (UID: \"239cfe46-3a3a-4624-9aa2-ea7112f28947\") " pod="openshift-marketplace/certified-operators-hcvm5" Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.332177 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:25 crc kubenswrapper[4948]: E0312 00:10:25.332455 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 00:10:25.83244268 +0000 UTC m=+225.288046418 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g4kcb" (UID: "af75ce0a-d6bb-48db-a422-08b1d87e1d3f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.345337 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7bb5cdccc6-pwdcd" event={"ID":"cd346d37-ec14-49f0-8e88-f73930b8bab8","Type":"ContainerStarted","Data":"e52800c551192345b03ce2fcafba9af613b213a83f507e64b193ef7ccd50b11e"} Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.345374 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7bb5cdccc6-pwdcd" event={"ID":"cd346d37-ec14-49f0-8e88-f73930b8bab8","Type":"ContainerStarted","Data":"4ef442baa04f491a58e34e7e726e281ce5c5a6d127017aca94eaf63dceefc1f0"} Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.346284 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7bb5cdccc6-pwdcd" Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.351073 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-62wnf" Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.356052 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-mqmtd" event={"ID":"be985ff1-de8a-431f-a230-67894ebe52d5","Type":"ContainerStarted","Data":"d40378a69a84913598b23646cad20a1b3bb8e519cacd031574948aa64abb9065"} Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.356088 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/network-metrics-daemon-mqmtd" event={"ID":"be985ff1-de8a-431f-a230-67894ebe52d5","Type":"ContainerStarted","Data":"25ee4f2411923b37a41bac6bdfb1522ecb20beeaf47490cbb1916953b12d327c"} Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.385822 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7bb5cdccc6-pwdcd" podStartSLOduration=4.385803467 podStartE2EDuration="4.385803467s" podCreationTimestamp="2026-03-12 00:10:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:10:25.382659094 +0000 UTC m=+224.838262822" watchObservedRunningTime="2026-03-12 00:10:25.385803467 +0000 UTC m=+224.841407205" Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.386141 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-network-console/networking-console-plugin-85b44fc459-gdk6g" event={"ID":"5fe485a1-e14f-4c09-b5b9-f252bc42b7e8","Type":"ContainerStarted","Data":"98592730a79592864ba0661f9e998534c11f3611f94e65464a5ccb8e2517a52b"} Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.414249 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-5q9rs" event={"ID":"4ba12474-c71f-429c-bcdf-9d74eb0b4b7a","Type":"ContainerStarted","Data":"4fe19962850420a5de8ab6ec047f18af96daa457f79eff5fb5fe179801d2ae8b"} Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.414515 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-5q9rs" event={"ID":"4ba12474-c71f-429c-bcdf-9d74eb0b4b7a","Type":"ContainerStarted","Data":"690210f622d723f1b2ee4fc15c1b0211b2b41a7910eae1897e150152d9d5f395"} Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.437658 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7bb5cdccc6-pwdcd" Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.437989 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rp8sv" event={"ID":"60024b0b-f817-477f-83f1-e53cbc1e05b0","Type":"ContainerDied","Data":"4795f512d0d3efb3414bf29827a15719ed2468da83504acccf6d4089a5361df6"} Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.438007 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-6576b87f9c-rp8sv" Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.438030 4948 scope.go:117] "RemoveContainer" containerID="567bd37219624454e44e80dc908b61d93b5a4feeeecc47a009a525503103bc90" Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.438649 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 00:10:25 crc kubenswrapper[4948]: E0312 00:10:25.439276 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 00:10:25.939256426 +0000 UTC m=+225.394860164 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.439497 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-pk6rv\" (UniqueName: \"kubernetes.io/projected/239cfe46-3a3a-4624-9aa2-ea7112f28947-kube-api-access-pk6rv\") pod \"certified-operators-hcvm5\" (UID: \"239cfe46-3a3a-4624-9aa2-ea7112f28947\") " pod="openshift-marketplace/certified-operators-hcvm5" Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.439529 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.439830 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/239cfe46-3a3a-4624-9aa2-ea7112f28947-utilities\") pod \"certified-operators-hcvm5\" (UID: \"239cfe46-3a3a-4624-9aa2-ea7112f28947\") " pod="openshift-marketplace/certified-operators-hcvm5" Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.440149 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/239cfe46-3a3a-4624-9aa2-ea7112f28947-catalog-content\") pod \"certified-operators-hcvm5\" (UID: \"239cfe46-3a3a-4624-9aa2-ea7112f28947\") " pod="openshift-marketplace/certified-operators-hcvm5" Mar 12 00:10:25 crc kubenswrapper[4948]: E0312 00:10:25.441420 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName: nodeName:}" failed. No retries permitted until 2026-03-12 00:10:25.941411657 +0000 UTC m=+225.397015395 (durationBeforeRetry 500ms). Error: MountVolume.MountDevice failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "image-registry-697d97f7c8-g4kcb" (UID: "af75ce0a-d6bb-48db-a422-08b1d87e1d3f") : kubernetes.io/csi: attacher.MountDevice failed to create newCsiDriverClient: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.442067 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/239cfe46-3a3a-4624-9aa2-ea7112f28947-utilities\") pod \"certified-operators-hcvm5\" (UID: \"239cfe46-3a3a-4624-9aa2-ea7112f28947\") " pod="openshift-marketplace/certified-operators-hcvm5" Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.442246 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/239cfe46-3a3a-4624-9aa2-ea7112f28947-catalog-content\") pod \"certified-operators-hcvm5\" (UID: \"239cfe46-3a3a-4624-9aa2-ea7112f28947\") " pod="openshift-marketplace/certified-operators-hcvm5" Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.460062 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-multus/network-metrics-daemon-mqmtd" podStartSLOduration=186.460042963 podStartE2EDuration="3m6.460042963s" podCreationTimestamp="2026-03-12 00:07:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:10:25.411832486 +0000 UTC m=+224.867436224" watchObservedRunningTime="2026-03-12 00:10:25.460042963 +0000 UTC m=+224.915646691" Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.524203 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-pk6rv\" (UniqueName: \"kubernetes.io/projected/239cfe46-3a3a-4624-9aa2-ea7112f28947-kube-api-access-pk6rv\") pod \"certified-operators-hcvm5\" (UID: \"239cfe46-3a3a-4624-9aa2-ea7112f28947\") " pod="openshift-marketplace/certified-operators-hcvm5" Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.524605 4948 generic.go:334] "Generic (PLEG): container finished" podID="5b801839-e39a-48a9-9eee-1bb95df5fbfb" containerID="542707e4c866aa968cbd0962bee2140bff37118c8f12f28bc2e587b0cd01b2d5" exitCode=0 Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.526306 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29554560-8mgv9" event={"ID":"5b801839-e39a-48a9-9eee-1bb95df5fbfb","Type":"ContainerDied","Data":"542707e4c866aa968cbd0962bee2140bff37118c8f12f28bc2e587b0cd01b2d5"} Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.541800 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 00:10:25 crc kubenswrapper[4948]: E0312 00:10:25.542355 4948 nestedpendingoperations.go:348] Operation for "{volumeName:kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8 podName:8f668bae-612b-4b75-9490-919e737c6a3b nodeName:}" failed. No retries permitted until 2026-03-12 00:10:26.042339887 +0000 UTC m=+225.497943615 (durationBeforeRetry 500ms). Error: UnmountVolume.TearDown failed for volume "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (UniqueName: "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b") : kubernetes.io/csi: Unmounter.TearDownAt failed to get CSI client: driver name kubevirt.io.hostpath-provisioner not found in the list of registered CSI drivers Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.561763 4948 reconciler.go:161] "OperationExecutor.RegisterPlugin started" plugin={"SocketPath":"/var/lib/kubelet/plugins_registry/kubevirt.io.hostpath-provisioner-reg.sock","Timestamp":"2026-03-12T00:10:24.934540928Z","Handler":null,"Name":""} Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.580542 4948 csi_plugin.go:100] kubernetes.io/csi: Trying to validate a new CSI Driver with name: kubevirt.io.hostpath-provisioner endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock versions: 1.0.0 Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.580576 4948 csi_plugin.go:113] kubernetes.io/csi: Register new plugin with name: kubevirt.io.hostpath-provisioner at endpoint: /var/lib/kubelet/plugins/csi-hostpath/csi.sock Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.603456 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-rp8sv"] Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.607073 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-6576b87f9c-rp8sv"] Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.643607 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.652930 4948 csi_attacher.go:380] kubernetes.io/csi: attacher.MountDevice STAGE_UNSTAGE_VOLUME capability not set. Skipping MountDevice... Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.652966 4948 operation_generator.go:580] "MountVolume.MountDevice succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") device mount path \"/var/lib/kubelet/plugins/kubernetes.io/csi/kubevirt.io.hostpath-provisioner/1f4776af88835e41c12b831b4c9fed40233456d14189815a54dbe7f892fc1983/globalmount\"" pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.692377 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hcvm5" Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.722828 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-697d97f7c8-g4kcb\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.747848 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"8f668bae-612b-4b75-9490-919e737c6a3b\" (UID: \"8f668bae-612b-4b75-9490-919e737c6a3b\") " Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.783296 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8") pod "8f668bae-612b-4b75-9490-919e737c6a3b" (UID: "8f668bae-612b-4b75-9490-919e737c6a3b"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.801270 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-rflhd"] Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.923040 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-hcvm5"] Mar 12 00:10:25 crc kubenswrapper[4948]: I0312 00:10:25.935600 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.001733 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-62wnf"] Mar 12 00:10:26 crc kubenswrapper[4948]: W0312 00:10:26.070696 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode6318b61_ef2d_41da_8c92_ff7523ce613f.slice/crio-470ffe54e5d1194c3f772e05a00e8e8d3d874f5fac0e64b8ffad2c49d79ba71f WatchSource:0}: Error finding container 470ffe54e5d1194c3f772e05a00e8e8d3d874f5fac0e64b8ffad2c49d79ba71f: Status 404 returned error can't find the container with id 470ffe54e5d1194c3f772e05a00e8e8d3d874f5fac0e64b8ffad2c49d79ba71f Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.096873 4948 patch_prober.go:28] interesting pod/router-default-5444994796-7vhkr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 12 00:10:26 crc kubenswrapper[4948]: [-]has-synced failed: reason withheld Mar 12 00:10:26 crc kubenswrapper[4948]: [+]process-running ok Mar 12 00:10:26 crc kubenswrapper[4948]: healthz check failed Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.096940 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7vhkr" podUID="e6184ca8-dc6e-4d34-ad0e-87c7107acb82" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.137477 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-g4kcb"] Mar 12 00:10:26 crc kubenswrapper[4948]: W0312 00:10:26.182014 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podaf75ce0a_d6bb_48db_a422_08b1d87e1d3f.slice/crio-58ba7b108182e3732e9b992a4f96564f36fb0731b067233c98e570d33cb87f51 WatchSource:0}: Error finding container 58ba7b108182e3732e9b992a4f96564f36fb0731b067233c98e570d33cb87f51: Status 404 returned error can't find the container with id 58ba7b108182e3732e9b992a4f96564f36fb0731b067233c98e570d33cb87f51 Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.223794 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-778c687499-dcjm2"] Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.224899 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-778c687499-dcjm2" Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.232170 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.232375 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.232378 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.232488 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.232616 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.232834 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.249852 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-778c687499-dcjm2"] Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.254902 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f87a342-cf29-45fe-acf3-2bfba03bea24-serving-cert\") pod \"route-controller-manager-778c687499-dcjm2\" (UID: \"8f87a342-cf29-45fe-acf3-2bfba03bea24\") " pod="openshift-route-controller-manager/route-controller-manager-778c687499-dcjm2" Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.254950 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-kgjb7\" (UniqueName: \"kubernetes.io/projected/8f87a342-cf29-45fe-acf3-2bfba03bea24-kube-api-access-kgjb7\") pod \"route-controller-manager-778c687499-dcjm2\" (UID: \"8f87a342-cf29-45fe-acf3-2bfba03bea24\") " pod="openshift-route-controller-manager/route-controller-manager-778c687499-dcjm2" Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.254983 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f87a342-cf29-45fe-acf3-2bfba03bea24-config\") pod \"route-controller-manager-778c687499-dcjm2\" (UID: \"8f87a342-cf29-45fe-acf3-2bfba03bea24\") " pod="openshift-route-controller-manager/route-controller-manager-778c687499-dcjm2" Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.255220 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8f87a342-cf29-45fe-acf3-2bfba03bea24-client-ca\") pod \"route-controller-manager-778c687499-dcjm2\" (UID: \"8f87a342-cf29-45fe-acf3-2bfba03bea24\") " pod="openshift-route-controller-manager/route-controller-manager-778c687499-dcjm2" Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.357100 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f87a342-cf29-45fe-acf3-2bfba03bea24-config\") pod \"route-controller-manager-778c687499-dcjm2\" (UID: \"8f87a342-cf29-45fe-acf3-2bfba03bea24\") " pod="openshift-route-controller-manager/route-controller-manager-778c687499-dcjm2" Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.357718 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8f87a342-cf29-45fe-acf3-2bfba03bea24-client-ca\") pod \"route-controller-manager-778c687499-dcjm2\" (UID: \"8f87a342-cf29-45fe-acf3-2bfba03bea24\") " pod="openshift-route-controller-manager/route-controller-manager-778c687499-dcjm2" Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.357750 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f87a342-cf29-45fe-acf3-2bfba03bea24-serving-cert\") pod \"route-controller-manager-778c687499-dcjm2\" (UID: \"8f87a342-cf29-45fe-acf3-2bfba03bea24\") " pod="openshift-route-controller-manager/route-controller-manager-778c687499-dcjm2" Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.357773 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-kgjb7\" (UniqueName: \"kubernetes.io/projected/8f87a342-cf29-45fe-acf3-2bfba03bea24-kube-api-access-kgjb7\") pod \"route-controller-manager-778c687499-dcjm2\" (UID: \"8f87a342-cf29-45fe-acf3-2bfba03bea24\") " pod="openshift-route-controller-manager/route-controller-manager-778c687499-dcjm2" Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.359978 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8f87a342-cf29-45fe-acf3-2bfba03bea24-client-ca\") pod \"route-controller-manager-778c687499-dcjm2\" (UID: \"8f87a342-cf29-45fe-acf3-2bfba03bea24\") " pod="openshift-route-controller-manager/route-controller-manager-778c687499-dcjm2" Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.362284 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f87a342-cf29-45fe-acf3-2bfba03bea24-config\") pod \"route-controller-manager-778c687499-dcjm2\" (UID: \"8f87a342-cf29-45fe-acf3-2bfba03bea24\") " pod="openshift-route-controller-manager/route-controller-manager-778c687499-dcjm2" Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.368258 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f87a342-cf29-45fe-acf3-2bfba03bea24-serving-cert\") pod \"route-controller-manager-778c687499-dcjm2\" (UID: \"8f87a342-cf29-45fe-acf3-2bfba03bea24\") " pod="openshift-route-controller-manager/route-controller-manager-778c687499-dcjm2" Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.374704 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-kgjb7\" (UniqueName: \"kubernetes.io/projected/8f87a342-cf29-45fe-acf3-2bfba03bea24-kube-api-access-kgjb7\") pod \"route-controller-manager-778c687499-dcjm2\" (UID: \"8f87a342-cf29-45fe-acf3-2bfba03bea24\") " pod="openshift-route-controller-manager/route-controller-manager-778c687499-dcjm2" Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.390504 4948 ???:1] "http: TLS handshake error from 192.168.126.11:35552: no serving certificate available for the kubelet" Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.425600 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-778c687499-dcjm2" Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.530797 4948 generic.go:334] "Generic (PLEG): container finished" podID="239cfe46-3a3a-4624-9aa2-ea7112f28947" containerID="9bad20e9704e9652c189e2d25b211eb7431aaf5b1083932c44e8b9e48b7c1dc7" exitCode=0 Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.530866 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hcvm5" event={"ID":"239cfe46-3a3a-4624-9aa2-ea7112f28947","Type":"ContainerDied","Data":"9bad20e9704e9652c189e2d25b211eb7431aaf5b1083932c44e8b9e48b7c1dc7"} Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.530911 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hcvm5" event={"ID":"239cfe46-3a3a-4624-9aa2-ea7112f28947","Type":"ContainerStarted","Data":"0cf8f970926697cdfac352fdf93f8628843806a504f3e5fcb5654e9560c5011d"} Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.534468 4948 generic.go:334] "Generic (PLEG): container finished" podID="5d83d568-3968-4ea6-b548-511c951defa5" containerID="9cece2d83a0870b294348ea430867de0e80ef150fd12302e1925a21396eb90d1" exitCode=0 Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.535045 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v8h88" event={"ID":"5d83d568-3968-4ea6-b548-511c951defa5","Type":"ContainerDied","Data":"9cece2d83a0870b294348ea430867de0e80ef150fd12302e1925a21396eb90d1"} Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.535075 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v8h88" event={"ID":"5d83d568-3968-4ea6-b548-511c951defa5","Type":"ContainerStarted","Data":"a6dc12e7d1d52b0588ce8348ac2a6750d029729d64b7ddce4efc6ab839856f2e"} Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.536542 4948 generic.go:334] "Generic (PLEG): container finished" podID="e6318b61-ef2d-41da-8c92-ff7523ce613f" containerID="bfa1f11178206f64a1411b6b67c539f5cd703dddb5b89197ebf8a1b95ff74ebd" exitCode=0 Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.536583 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-62wnf" event={"ID":"e6318b61-ef2d-41da-8c92-ff7523ce613f","Type":"ContainerDied","Data":"bfa1f11178206f64a1411b6b67c539f5cd703dddb5b89197ebf8a1b95ff74ebd"} Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.536598 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-62wnf" event={"ID":"e6318b61-ef2d-41da-8c92-ff7523ce613f","Type":"ContainerStarted","Data":"470ffe54e5d1194c3f772e05a00e8e8d3d874f5fac0e64b8ffad2c49d79ba71f"} Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.540036 4948 generic.go:334] "Generic (PLEG): container finished" podID="5184125a-dde1-4db4-9d6b-5b3c45d8d82a" containerID="f428f193760fb0cfa943b3d75ca100ec3dc34883d4ebe517835fa46ecb35b01e" exitCode=0 Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.540457 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rflhd" event={"ID":"5184125a-dde1-4db4-9d6b-5b3c45d8d82a","Type":"ContainerDied","Data":"f428f193760fb0cfa943b3d75ca100ec3dc34883d4ebe517835fa46ecb35b01e"} Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.540477 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rflhd" event={"ID":"5184125a-dde1-4db4-9d6b-5b3c45d8d82a","Type":"ContainerStarted","Data":"68d6d263e322261e4b32d31022aad0d2424639725173f965e7a064ec610c70e0"} Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.565586 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" event={"ID":"af75ce0a-d6bb-48db-a422-08b1d87e1d3f","Type":"ContainerStarted","Data":"3870a7ecb039300e799b9a72a8748670d0c4b779f0a6487f32b8a103b3d173ec"} Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.565659 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" event={"ID":"af75ce0a-d6bb-48db-a422-08b1d87e1d3f","Type":"ContainerStarted","Data":"58ba7b108182e3732e9b992a4f96564f36fb0731b067233c98e570d33cb87f51"} Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.565695 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.569659 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="hostpath-provisioner/csi-hostpathplugin-5q9rs" event={"ID":"4ba12474-c71f-429c-bcdf-9d74eb0b4b7a","Type":"ContainerStarted","Data":"0359afbe752165730676e14871906cc897525d3894784fc8ef9e19f979c1f202"} Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.637188 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" podStartSLOduration=187.637167381 podStartE2EDuration="3m7.637167381s" podCreationTimestamp="2026-03-12 00:07:19 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:10:26.635131484 +0000 UTC m=+226.090735232" watchObservedRunningTime="2026-03-12 00:10:26.637167381 +0000 UTC m=+226.092771119" Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.706621 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-778c687499-dcjm2"] Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.838699 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="hostpath-provisioner/csi-hostpathplugin-5q9rs" podStartSLOduration=11.838679042 podStartE2EDuration="11.838679042s" podCreationTimestamp="2026-03-12 00:10:15 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:10:26.702422097 +0000 UTC m=+226.158025835" watchObservedRunningTime="2026-03-12 00:10:26.838679042 +0000 UTC m=+226.294282780" Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.839245 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-zc9zh"] Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.839967 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29554560-8mgv9" Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.840519 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zc9zh" Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.846179 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.847303 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zc9zh"] Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.893343 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5b801839-e39a-48a9-9eee-1bb95df5fbfb-secret-volume\") pod \"5b801839-e39a-48a9-9eee-1bb95df5fbfb\" (UID: \"5b801839-e39a-48a9-9eee-1bb95df5fbfb\") " Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.893396 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5b801839-e39a-48a9-9eee-1bb95df5fbfb-config-volume\") pod \"5b801839-e39a-48a9-9eee-1bb95df5fbfb\" (UID: \"5b801839-e39a-48a9-9eee-1bb95df5fbfb\") " Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.893437 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fzbkp\" (UniqueName: \"kubernetes.io/projected/5b801839-e39a-48a9-9eee-1bb95df5fbfb-kube-api-access-fzbkp\") pod \"5b801839-e39a-48a9-9eee-1bb95df5fbfb\" (UID: \"5b801839-e39a-48a9-9eee-1bb95df5fbfb\") " Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.893612 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tck76\" (UniqueName: \"kubernetes.io/projected/f8cf7ee1-e662-4afd-9e08-112171677c22-kube-api-access-tck76\") pod \"redhat-marketplace-zc9zh\" (UID: \"f8cf7ee1-e662-4afd-9e08-112171677c22\") " pod="openshift-marketplace/redhat-marketplace-zc9zh" Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.893651 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8cf7ee1-e662-4afd-9e08-112171677c22-utilities\") pod \"redhat-marketplace-zc9zh\" (UID: \"f8cf7ee1-e662-4afd-9e08-112171677c22\") " pod="openshift-marketplace/redhat-marketplace-zc9zh" Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.893678 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8cf7ee1-e662-4afd-9e08-112171677c22-catalog-content\") pod \"redhat-marketplace-zc9zh\" (UID: \"f8cf7ee1-e662-4afd-9e08-112171677c22\") " pod="openshift-marketplace/redhat-marketplace-zc9zh" Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.898705 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5b801839-e39a-48a9-9eee-1bb95df5fbfb-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "5b801839-e39a-48a9-9eee-1bb95df5fbfb" (UID: "5b801839-e39a-48a9-9eee-1bb95df5fbfb"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.898783 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5b801839-e39a-48a9-9eee-1bb95df5fbfb-config-volume" (OuterVolumeSpecName: "config-volume") pod "5b801839-e39a-48a9-9eee-1bb95df5fbfb" (UID: "5b801839-e39a-48a9-9eee-1bb95df5fbfb"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.899646 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5b801839-e39a-48a9-9eee-1bb95df5fbfb-kube-api-access-fzbkp" (OuterVolumeSpecName: "kube-api-access-fzbkp") pod "5b801839-e39a-48a9-9eee-1bb95df5fbfb" (UID: "5b801839-e39a-48a9-9eee-1bb95df5fbfb"). InnerVolumeSpecName "kube-api-access-fzbkp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.994892 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8cf7ee1-e662-4afd-9e08-112171677c22-utilities\") pod \"redhat-marketplace-zc9zh\" (UID: \"f8cf7ee1-e662-4afd-9e08-112171677c22\") " pod="openshift-marketplace/redhat-marketplace-zc9zh" Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.994964 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8cf7ee1-e662-4afd-9e08-112171677c22-catalog-content\") pod \"redhat-marketplace-zc9zh\" (UID: \"f8cf7ee1-e662-4afd-9e08-112171677c22\") " pod="openshift-marketplace/redhat-marketplace-zc9zh" Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.995044 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tck76\" (UniqueName: \"kubernetes.io/projected/f8cf7ee1-e662-4afd-9e08-112171677c22-kube-api-access-tck76\") pod \"redhat-marketplace-zc9zh\" (UID: \"f8cf7ee1-e662-4afd-9e08-112171677c22\") " pod="openshift-marketplace/redhat-marketplace-zc9zh" Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.995087 4948 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5b801839-e39a-48a9-9eee-1bb95df5fbfb-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.995102 4948 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5b801839-e39a-48a9-9eee-1bb95df5fbfb-config-volume\") on node \"crc\" DevicePath \"\"" Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.995114 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fzbkp\" (UniqueName: \"kubernetes.io/projected/5b801839-e39a-48a9-9eee-1bb95df5fbfb-kube-api-access-fzbkp\") on node \"crc\" DevicePath \"\"" Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.995911 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8cf7ee1-e662-4afd-9e08-112171677c22-utilities\") pod \"redhat-marketplace-zc9zh\" (UID: \"f8cf7ee1-e662-4afd-9e08-112171677c22\") " pod="openshift-marketplace/redhat-marketplace-zc9zh" Mar 12 00:10:26 crc kubenswrapper[4948]: I0312 00:10:26.996199 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8cf7ee1-e662-4afd-9e08-112171677c22-catalog-content\") pod \"redhat-marketplace-zc9zh\" (UID: \"f8cf7ee1-e662-4afd-9e08-112171677c22\") " pod="openshift-marketplace/redhat-marketplace-zc9zh" Mar 12 00:10:27 crc kubenswrapper[4948]: I0312 00:10:27.010775 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tck76\" (UniqueName: \"kubernetes.io/projected/f8cf7ee1-e662-4afd-9e08-112171677c22-kube-api-access-tck76\") pod \"redhat-marketplace-zc9zh\" (UID: \"f8cf7ee1-e662-4afd-9e08-112171677c22\") " pod="openshift-marketplace/redhat-marketplace-zc9zh" Mar 12 00:10:27 crc kubenswrapper[4948]: I0312 00:10:27.095939 4948 patch_prober.go:28] interesting pod/router-default-5444994796-7vhkr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 12 00:10:27 crc kubenswrapper[4948]: [-]has-synced failed: reason withheld Mar 12 00:10:27 crc kubenswrapper[4948]: [+]process-running ok Mar 12 00:10:27 crc kubenswrapper[4948]: healthz check failed Mar 12 00:10:27 crc kubenswrapper[4948]: I0312 00:10:27.096002 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7vhkr" podUID="e6184ca8-dc6e-4d34-ad0e-87c7107acb82" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 12 00:10:27 crc kubenswrapper[4948]: I0312 00:10:27.162046 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zc9zh" Mar 12 00:10:27 crc kubenswrapper[4948]: I0312 00:10:27.232618 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-bsmcp"] Mar 12 00:10:27 crc kubenswrapper[4948]: E0312 00:10:27.235099 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5b801839-e39a-48a9-9eee-1bb95df5fbfb" containerName="collect-profiles" Mar 12 00:10:27 crc kubenswrapper[4948]: I0312 00:10:27.235122 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="5b801839-e39a-48a9-9eee-1bb95df5fbfb" containerName="collect-profiles" Mar 12 00:10:27 crc kubenswrapper[4948]: I0312 00:10:27.235256 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="5b801839-e39a-48a9-9eee-1bb95df5fbfb" containerName="collect-profiles" Mar 12 00:10:27 crc kubenswrapper[4948]: I0312 00:10:27.236269 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bsmcp" Mar 12 00:10:27 crc kubenswrapper[4948]: I0312 00:10:27.241621 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bsmcp"] Mar 12 00:10:27 crc kubenswrapper[4948]: I0312 00:10:27.299739 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44212e8e-c675-4643-a752-997be5f8e980-catalog-content\") pod \"redhat-marketplace-bsmcp\" (UID: \"44212e8e-c675-4643-a752-997be5f8e980\") " pod="openshift-marketplace/redhat-marketplace-bsmcp" Mar 12 00:10:27 crc kubenswrapper[4948]: I0312 00:10:27.299828 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-h267n\" (UniqueName: \"kubernetes.io/projected/44212e8e-c675-4643-a752-997be5f8e980-kube-api-access-h267n\") pod \"redhat-marketplace-bsmcp\" (UID: \"44212e8e-c675-4643-a752-997be5f8e980\") " pod="openshift-marketplace/redhat-marketplace-bsmcp" Mar 12 00:10:27 crc kubenswrapper[4948]: I0312 00:10:27.299914 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44212e8e-c675-4643-a752-997be5f8e980-utilities\") pod \"redhat-marketplace-bsmcp\" (UID: \"44212e8e-c675-4643-a752-997be5f8e980\") " pod="openshift-marketplace/redhat-marketplace-bsmcp" Mar 12 00:10:27 crc kubenswrapper[4948]: I0312 00:10:27.378739 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="60024b0b-f817-477f-83f1-e53cbc1e05b0" path="/var/lib/kubelet/pods/60024b0b-f817-477f-83f1-e53cbc1e05b0/volumes" Mar 12 00:10:27 crc kubenswrapper[4948]: I0312 00:10:27.380050 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f668bae-612b-4b75-9490-919e737c6a3b" path="/var/lib/kubelet/pods/8f668bae-612b-4b75-9490-919e737c6a3b/volumes" Mar 12 00:10:27 crc kubenswrapper[4948]: I0312 00:10:27.380718 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-apiserver/apiserver-76f77b778f-6jfc4" Mar 12 00:10:27 crc kubenswrapper[4948]: I0312 00:10:27.385556 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-apiserver/apiserver-76f77b778f-6jfc4" Mar 12 00:10:27 crc kubenswrapper[4948]: I0312 00:10:27.401438 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44212e8e-c675-4643-a752-997be5f8e980-utilities\") pod \"redhat-marketplace-bsmcp\" (UID: \"44212e8e-c675-4643-a752-997be5f8e980\") " pod="openshift-marketplace/redhat-marketplace-bsmcp" Mar 12 00:10:27 crc kubenswrapper[4948]: I0312 00:10:27.401489 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44212e8e-c675-4643-a752-997be5f8e980-catalog-content\") pod \"redhat-marketplace-bsmcp\" (UID: \"44212e8e-c675-4643-a752-997be5f8e980\") " pod="openshift-marketplace/redhat-marketplace-bsmcp" Mar 12 00:10:27 crc kubenswrapper[4948]: I0312 00:10:27.401512 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-h267n\" (UniqueName: \"kubernetes.io/projected/44212e8e-c675-4643-a752-997be5f8e980-kube-api-access-h267n\") pod \"redhat-marketplace-bsmcp\" (UID: \"44212e8e-c675-4643-a752-997be5f8e980\") " pod="openshift-marketplace/redhat-marketplace-bsmcp" Mar 12 00:10:27 crc kubenswrapper[4948]: I0312 00:10:27.402016 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44212e8e-c675-4643-a752-997be5f8e980-utilities\") pod \"redhat-marketplace-bsmcp\" (UID: \"44212e8e-c675-4643-a752-997be5f8e980\") " pod="openshift-marketplace/redhat-marketplace-bsmcp" Mar 12 00:10:27 crc kubenswrapper[4948]: I0312 00:10:27.402071 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44212e8e-c675-4643-a752-997be5f8e980-catalog-content\") pod \"redhat-marketplace-bsmcp\" (UID: \"44212e8e-c675-4643-a752-997be5f8e980\") " pod="openshift-marketplace/redhat-marketplace-bsmcp" Mar 12 00:10:27 crc kubenswrapper[4948]: I0312 00:10:27.461084 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-h267n\" (UniqueName: \"kubernetes.io/projected/44212e8e-c675-4643-a752-997be5f8e980-kube-api-access-h267n\") pod \"redhat-marketplace-bsmcp\" (UID: \"44212e8e-c675-4643-a752-997be5f8e980\") " pod="openshift-marketplace/redhat-marketplace-bsmcp" Mar 12 00:10:27 crc kubenswrapper[4948]: I0312 00:10:27.553316 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-zc9zh"] Mar 12 00:10:27 crc kubenswrapper[4948]: I0312 00:10:27.574037 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bsmcp" Mar 12 00:10:27 crc kubenswrapper[4948]: I0312 00:10:27.622744 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29554560-8mgv9" event={"ID":"5b801839-e39a-48a9-9eee-1bb95df5fbfb","Type":"ContainerDied","Data":"a043eedcdc4b607d199f5e715c5a48cdc1eac9b39adc704584885494a0fa2ab7"} Mar 12 00:10:27 crc kubenswrapper[4948]: I0312 00:10:27.622782 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a043eedcdc4b607d199f5e715c5a48cdc1eac9b39adc704584885494a0fa2ab7" Mar 12 00:10:27 crc kubenswrapper[4948]: I0312 00:10:27.622831 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29554560-8mgv9" Mar 12 00:10:27 crc kubenswrapper[4948]: I0312 00:10:27.631409 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-778c687499-dcjm2" event={"ID":"8f87a342-cf29-45fe-acf3-2bfba03bea24","Type":"ContainerStarted","Data":"778babe80ddb79b698b590dae2eae24873266a62a5bdc2c6418e23eaa451fea0"} Mar 12 00:10:27 crc kubenswrapper[4948]: I0312 00:10:27.631453 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-778c687499-dcjm2" event={"ID":"8f87a342-cf29-45fe-acf3-2bfba03bea24","Type":"ContainerStarted","Data":"b9b1a41b07b3648b422714ea2b8280dcdbf2062fc282b54a5bf8486d324cc6e2"} Mar 12 00:10:27 crc kubenswrapper[4948]: I0312 00:10:27.631883 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-778c687499-dcjm2" Mar 12 00:10:27 crc kubenswrapper[4948]: I0312 00:10:27.644168 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-778c687499-dcjm2" Mar 12 00:10:27 crc kubenswrapper[4948]: I0312 00:10:27.669993 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-778c687499-dcjm2" podStartSLOduration=5.669976216 podStartE2EDuration="5.669976216s" podCreationTimestamp="2026-03-12 00:10:22 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:10:27.667979519 +0000 UTC m=+227.123583257" watchObservedRunningTime="2026-03-12 00:10:27.669976216 +0000 UTC m=+227.125579954" Mar 12 00:10:27 crc kubenswrapper[4948]: I0312 00:10:27.731832 4948 patch_prober.go:28] interesting pod/downloads-7954f5f757-xdh6w container/download-server namespace/openshift-console: Liveness probe status=failure output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" start-of-body= Mar 12 00:10:27 crc kubenswrapper[4948]: I0312 00:10:27.732223 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-console/downloads-7954f5f757-xdh6w" podUID="74ab81d8-55e8-47da-adb5-1f616a100a31" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" Mar 12 00:10:27 crc kubenswrapper[4948]: I0312 00:10:27.732562 4948 patch_prober.go:28] interesting pod/downloads-7954f5f757-xdh6w container/download-server namespace/openshift-console: Readiness probe status=failure output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" start-of-body= Mar 12 00:10:27 crc kubenswrapper[4948]: I0312 00:10:27.732584 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-console/downloads-7954f5f757-xdh6w" podUID="74ab81d8-55e8-47da-adb5-1f616a100a31" containerName="download-server" probeResult="failure" output="Get \"http://10.217.0.16:8080/\": dial tcp 10.217.0.16:8080: connect: connection refused" Mar 12 00:10:27 crc kubenswrapper[4948]: I0312 00:10:27.762496 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-console/console-f9d7485db-qq8gg" Mar 12 00:10:27 crc kubenswrapper[4948]: I0312 00:10:27.762559 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-console/console-f9d7485db-qq8gg" Mar 12 00:10:27 crc kubenswrapper[4948]: I0312 00:10:27.768638 4948 patch_prober.go:28] interesting pod/console-f9d7485db-qq8gg container/console namespace/openshift-console: Startup probe status=failure output="Get \"https://10.217.0.15:8443/health\": dial tcp 10.217.0.15:8443: connect: connection refused" start-of-body= Mar 12 00:10:27 crc kubenswrapper[4948]: I0312 00:10:27.768687 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-console/console-f9d7485db-qq8gg" podUID="892017c6-01a1-4798-9d23-fff4f44c6932" containerName="console" probeResult="failure" output="Get \"https://10.217.0.15:8443/health\": dial tcp 10.217.0.15:8443: connect: connection refused" Mar 12 00:10:27 crc kubenswrapper[4948]: I0312 00:10:27.841824 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-btd4j"] Mar 12 00:10:27 crc kubenswrapper[4948]: I0312 00:10:27.847257 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-btd4j" Mar 12 00:10:27 crc kubenswrapper[4948]: I0312 00:10:27.855709 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 12 00:10:27 crc kubenswrapper[4948]: I0312 00:10:27.855716 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-btd4j"] Mar 12 00:10:27 crc kubenswrapper[4948]: I0312 00:10:27.858481 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-bsmcp"] Mar 12 00:10:27 crc kubenswrapper[4948]: I0312 00:10:27.914493 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-l7p47\" (UniqueName: \"kubernetes.io/projected/8a6ce79a-afdb-4e21-9e24-8de6924279bf-kube-api-access-l7p47\") pod \"redhat-operators-btd4j\" (UID: \"8a6ce79a-afdb-4e21-9e24-8de6924279bf\") " pod="openshift-marketplace/redhat-operators-btd4j" Mar 12 00:10:27 crc kubenswrapper[4948]: I0312 00:10:27.914572 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a6ce79a-afdb-4e21-9e24-8de6924279bf-catalog-content\") pod \"redhat-operators-btd4j\" (UID: \"8a6ce79a-afdb-4e21-9e24-8de6924279bf\") " pod="openshift-marketplace/redhat-operators-btd4j" Mar 12 00:10:27 crc kubenswrapper[4948]: I0312 00:10:27.914595 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a6ce79a-afdb-4e21-9e24-8de6924279bf-utilities\") pod \"redhat-operators-btd4j\" (UID: \"8a6ce79a-afdb-4e21-9e24-8de6924279bf\") " pod="openshift-marketplace/redhat-operators-btd4j" Mar 12 00:10:28 crc kubenswrapper[4948]: I0312 00:10:28.016076 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a6ce79a-afdb-4e21-9e24-8de6924279bf-catalog-content\") pod \"redhat-operators-btd4j\" (UID: \"8a6ce79a-afdb-4e21-9e24-8de6924279bf\") " pod="openshift-marketplace/redhat-operators-btd4j" Mar 12 00:10:28 crc kubenswrapper[4948]: I0312 00:10:28.016121 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a6ce79a-afdb-4e21-9e24-8de6924279bf-utilities\") pod \"redhat-operators-btd4j\" (UID: \"8a6ce79a-afdb-4e21-9e24-8de6924279bf\") " pod="openshift-marketplace/redhat-operators-btd4j" Mar 12 00:10:28 crc kubenswrapper[4948]: I0312 00:10:28.016172 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-l7p47\" (UniqueName: \"kubernetes.io/projected/8a6ce79a-afdb-4e21-9e24-8de6924279bf-kube-api-access-l7p47\") pod \"redhat-operators-btd4j\" (UID: \"8a6ce79a-afdb-4e21-9e24-8de6924279bf\") " pod="openshift-marketplace/redhat-operators-btd4j" Mar 12 00:10:28 crc kubenswrapper[4948]: I0312 00:10:28.016972 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a6ce79a-afdb-4e21-9e24-8de6924279bf-catalog-content\") pod \"redhat-operators-btd4j\" (UID: \"8a6ce79a-afdb-4e21-9e24-8de6924279bf\") " pod="openshift-marketplace/redhat-operators-btd4j" Mar 12 00:10:28 crc kubenswrapper[4948]: I0312 00:10:28.017173 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a6ce79a-afdb-4e21-9e24-8de6924279bf-utilities\") pod \"redhat-operators-btd4j\" (UID: \"8a6ce79a-afdb-4e21-9e24-8de6924279bf\") " pod="openshift-marketplace/redhat-operators-btd4j" Mar 12 00:10:28 crc kubenswrapper[4948]: I0312 00:10:28.032680 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-6w6fq"] Mar 12 00:10:28 crc kubenswrapper[4948]: I0312 00:10:28.038748 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6w6fq" Mar 12 00:10:28 crc kubenswrapper[4948]: I0312 00:10:28.050672 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-l7p47\" (UniqueName: \"kubernetes.io/projected/8a6ce79a-afdb-4e21-9e24-8de6924279bf-kube-api-access-l7p47\") pod \"redhat-operators-btd4j\" (UID: \"8a6ce79a-afdb-4e21-9e24-8de6924279bf\") " pod="openshift-marketplace/redhat-operators-btd4j" Mar 12 00:10:28 crc kubenswrapper[4948]: I0312 00:10:28.065633 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6w6fq"] Mar 12 00:10:28 crc kubenswrapper[4948]: I0312 00:10:28.093879 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ingress/router-default-5444994796-7vhkr" Mar 12 00:10:28 crc kubenswrapper[4948]: I0312 00:10:28.098404 4948 patch_prober.go:28] interesting pod/router-default-5444994796-7vhkr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 12 00:10:28 crc kubenswrapper[4948]: [-]has-synced failed: reason withheld Mar 12 00:10:28 crc kubenswrapper[4948]: [+]process-running ok Mar 12 00:10:28 crc kubenswrapper[4948]: healthz check failed Mar 12 00:10:28 crc kubenswrapper[4948]: I0312 00:10:28.098467 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7vhkr" podUID="e6184ca8-dc6e-4d34-ad0e-87c7107acb82" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 12 00:10:28 crc kubenswrapper[4948]: I0312 00:10:28.116900 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/529630b8-cc23-44b7-bafd-6ae1313877d1-utilities\") pod \"redhat-operators-6w6fq\" (UID: \"529630b8-cc23-44b7-bafd-6ae1313877d1\") " pod="openshift-marketplace/redhat-operators-6w6fq" Mar 12 00:10:28 crc kubenswrapper[4948]: I0312 00:10:28.116952 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-zx4xz\" (UniqueName: \"kubernetes.io/projected/529630b8-cc23-44b7-bafd-6ae1313877d1-kube-api-access-zx4xz\") pod \"redhat-operators-6w6fq\" (UID: \"529630b8-cc23-44b7-bafd-6ae1313877d1\") " pod="openshift-marketplace/redhat-operators-6w6fq" Mar 12 00:10:28 crc kubenswrapper[4948]: I0312 00:10:28.116989 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/529630b8-cc23-44b7-bafd-6ae1313877d1-catalog-content\") pod \"redhat-operators-6w6fq\" (UID: \"529630b8-cc23-44b7-bafd-6ae1313877d1\") " pod="openshift-marketplace/redhat-operators-6w6fq" Mar 12 00:10:28 crc kubenswrapper[4948]: I0312 00:10:28.213844 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-btd4j" Mar 12 00:10:28 crc kubenswrapper[4948]: I0312 00:10:28.217915 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/529630b8-cc23-44b7-bafd-6ae1313877d1-utilities\") pod \"redhat-operators-6w6fq\" (UID: \"529630b8-cc23-44b7-bafd-6ae1313877d1\") " pod="openshift-marketplace/redhat-operators-6w6fq" Mar 12 00:10:28 crc kubenswrapper[4948]: I0312 00:10:28.217979 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-zx4xz\" (UniqueName: \"kubernetes.io/projected/529630b8-cc23-44b7-bafd-6ae1313877d1-kube-api-access-zx4xz\") pod \"redhat-operators-6w6fq\" (UID: \"529630b8-cc23-44b7-bafd-6ae1313877d1\") " pod="openshift-marketplace/redhat-operators-6w6fq" Mar 12 00:10:28 crc kubenswrapper[4948]: I0312 00:10:28.218032 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/529630b8-cc23-44b7-bafd-6ae1313877d1-catalog-content\") pod \"redhat-operators-6w6fq\" (UID: \"529630b8-cc23-44b7-bafd-6ae1313877d1\") " pod="openshift-marketplace/redhat-operators-6w6fq" Mar 12 00:10:28 crc kubenswrapper[4948]: I0312 00:10:28.218410 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/529630b8-cc23-44b7-bafd-6ae1313877d1-utilities\") pod \"redhat-operators-6w6fq\" (UID: \"529630b8-cc23-44b7-bafd-6ae1313877d1\") " pod="openshift-marketplace/redhat-operators-6w6fq" Mar 12 00:10:28 crc kubenswrapper[4948]: I0312 00:10:28.218583 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/529630b8-cc23-44b7-bafd-6ae1313877d1-catalog-content\") pod \"redhat-operators-6w6fq\" (UID: \"529630b8-cc23-44b7-bafd-6ae1313877d1\") " pod="openshift-marketplace/redhat-operators-6w6fq" Mar 12 00:10:28 crc kubenswrapper[4948]: I0312 00:10:28.236047 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-zx4xz\" (UniqueName: \"kubernetes.io/projected/529630b8-cc23-44b7-bafd-6ae1313877d1-kube-api-access-zx4xz\") pod \"redhat-operators-6w6fq\" (UID: \"529630b8-cc23-44b7-bafd-6ae1313877d1\") " pod="openshift-marketplace/redhat-operators-6w6fq" Mar 12 00:10:28 crc kubenswrapper[4948]: I0312 00:10:28.286582 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 12 00:10:28 crc kubenswrapper[4948]: I0312 00:10:28.287260 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 12 00:10:28 crc kubenswrapper[4948]: I0312 00:10:28.291129 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Mar 12 00:10:28 crc kubenswrapper[4948]: I0312 00:10:28.291756 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Mar 12 00:10:28 crc kubenswrapper[4948]: I0312 00:10:28.303400 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 12 00:10:28 crc kubenswrapper[4948]: I0312 00:10:28.318681 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/02288ecc-afa1-48aa-9a2d-cc78303ede88-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"02288ecc-afa1-48aa-9a2d-cc78303ede88\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 12 00:10:28 crc kubenswrapper[4948]: I0312 00:10:28.318726 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/02288ecc-afa1-48aa-9a2d-cc78303ede88-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"02288ecc-afa1-48aa-9a2d-cc78303ede88\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 12 00:10:28 crc kubenswrapper[4948]: I0312 00:10:28.373442 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6w6fq" Mar 12 00:10:28 crc kubenswrapper[4948]: I0312 00:10:28.420123 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/02288ecc-afa1-48aa-9a2d-cc78303ede88-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"02288ecc-afa1-48aa-9a2d-cc78303ede88\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 12 00:10:28 crc kubenswrapper[4948]: I0312 00:10:28.420210 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/02288ecc-afa1-48aa-9a2d-cc78303ede88-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"02288ecc-afa1-48aa-9a2d-cc78303ede88\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 12 00:10:28 crc kubenswrapper[4948]: I0312 00:10:28.420265 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/02288ecc-afa1-48aa-9a2d-cc78303ede88-kubelet-dir\") pod \"revision-pruner-8-crc\" (UID: \"02288ecc-afa1-48aa-9a2d-cc78303ede88\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 12 00:10:28 crc kubenswrapper[4948]: I0312 00:10:28.442050 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/02288ecc-afa1-48aa-9a2d-cc78303ede88-kube-api-access\") pod \"revision-pruner-8-crc\" (UID: \"02288ecc-afa1-48aa-9a2d-cc78303ede88\") " pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 12 00:10:28 crc kubenswrapper[4948]: I0312 00:10:28.638277 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 12 00:10:28 crc kubenswrapper[4948]: I0312 00:10:28.661381 4948 generic.go:334] "Generic (PLEG): container finished" podID="44212e8e-c675-4643-a752-997be5f8e980" containerID="62ce39169fa39ff6e1b56e3fea7f604f7e9f4fd54c2e527988836d1a811b10e3" exitCode=0 Mar 12 00:10:28 crc kubenswrapper[4948]: I0312 00:10:28.661497 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bsmcp" event={"ID":"44212e8e-c675-4643-a752-997be5f8e980","Type":"ContainerDied","Data":"62ce39169fa39ff6e1b56e3fea7f604f7e9f4fd54c2e527988836d1a811b10e3"} Mar 12 00:10:28 crc kubenswrapper[4948]: I0312 00:10:28.661524 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bsmcp" event={"ID":"44212e8e-c675-4643-a752-997be5f8e980","Type":"ContainerStarted","Data":"05d255416a0db402af61232adca03cefca9928f3acf1a396523501205728b4a8"} Mar 12 00:10:28 crc kubenswrapper[4948]: I0312 00:10:28.665596 4948 generic.go:334] "Generic (PLEG): container finished" podID="f8cf7ee1-e662-4afd-9e08-112171677c22" containerID="385a4db639f3a271aeab550575db73d3ae75f07ef98cd979694d081c9ec6756e" exitCode=0 Mar 12 00:10:28 crc kubenswrapper[4948]: I0312 00:10:28.665645 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zc9zh" event={"ID":"f8cf7ee1-e662-4afd-9e08-112171677c22","Type":"ContainerDied","Data":"385a4db639f3a271aeab550575db73d3ae75f07ef98cd979694d081c9ec6756e"} Mar 12 00:10:28 crc kubenswrapper[4948]: I0312 00:10:28.665696 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zc9zh" event={"ID":"f8cf7ee1-e662-4afd-9e08-112171677c22","Type":"ContainerStarted","Data":"759cc547858b754c4e14a3f3cc4a8e48141ee9196217d84770c4d3cfa6e8e8c4"} Mar 12 00:10:28 crc kubenswrapper[4948]: I0312 00:10:28.716977 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 12 00:10:28 crc kubenswrapper[4948]: I0312 00:10:28.718134 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 12 00:10:28 crc kubenswrapper[4948]: I0312 00:10:28.720046 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager"/"installer-sa-dockercfg-kjl2n" Mar 12 00:10:28 crc kubenswrapper[4948]: I0312 00:10:28.721133 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager"/"kube-root-ca.crt" Mar 12 00:10:28 crc kubenswrapper[4948]: I0312 00:10:28.734177 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 12 00:10:28 crc kubenswrapper[4948]: I0312 00:10:28.754080 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-btd4j"] Mar 12 00:10:28 crc kubenswrapper[4948]: I0312 00:10:28.825587 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/44d3c29e-5c1e-41be-ae1a-8d1e5f9e39c7-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"44d3c29e-5c1e-41be-ae1a-8d1e5f9e39c7\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 12 00:10:28 crc kubenswrapper[4948]: I0312 00:10:28.826322 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/44d3c29e-5c1e-41be-ae1a-8d1e5f9e39c7-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"44d3c29e-5c1e-41be-ae1a-8d1e5f9e39c7\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 12 00:10:28 crc kubenswrapper[4948]: I0312 00:10:28.864698 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-6w6fq"] Mar 12 00:10:28 crc kubenswrapper[4948]: I0312 00:10:28.927827 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/44d3c29e-5c1e-41be-ae1a-8d1e5f9e39c7-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"44d3c29e-5c1e-41be-ae1a-8d1e5f9e39c7\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 12 00:10:28 crc kubenswrapper[4948]: I0312 00:10:28.927947 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/44d3c29e-5c1e-41be-ae1a-8d1e5f9e39c7-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"44d3c29e-5c1e-41be-ae1a-8d1e5f9e39c7\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 12 00:10:28 crc kubenswrapper[4948]: I0312 00:10:28.927957 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/44d3c29e-5c1e-41be-ae1a-8d1e5f9e39c7-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"44d3c29e-5c1e-41be-ae1a-8d1e5f9e39c7\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 12 00:10:28 crc kubenswrapper[4948]: I0312 00:10:28.948640 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/44d3c29e-5c1e-41be-ae1a-8d1e5f9e39c7-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"44d3c29e-5c1e-41be-ae1a-8d1e5f9e39c7\") " pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 12 00:10:29 crc kubenswrapper[4948]: I0312 00:10:29.042112 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 12 00:10:29 crc kubenswrapper[4948]: I0312 00:10:29.098150 4948 patch_prober.go:28] interesting pod/router-default-5444994796-7vhkr container/router namespace/openshift-ingress: Startup probe status=failure output="HTTP probe failed with statuscode: 500" start-of-body=[-]backend-http failed: reason withheld Mar 12 00:10:29 crc kubenswrapper[4948]: [-]has-synced failed: reason withheld Mar 12 00:10:29 crc kubenswrapper[4948]: [+]process-running ok Mar 12 00:10:29 crc kubenswrapper[4948]: healthz check failed Mar 12 00:10:29 crc kubenswrapper[4948]: I0312 00:10:29.098768 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-ingress/router-default-5444994796-7vhkr" podUID="e6184ca8-dc6e-4d34-ad0e-87c7107acb82" containerName="router" probeResult="failure" output="HTTP probe failed with statuscode: 500" Mar 12 00:10:29 crc kubenswrapper[4948]: I0312 00:10:29.224211 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-8-crc"] Mar 12 00:10:29 crc kubenswrapper[4948]: I0312 00:10:29.624275 4948 ???:1] "http: TLS handshake error from 192.168.126.11:37930: no serving certificate available for the kubelet" Mar 12 00:10:29 crc kubenswrapper[4948]: I0312 00:10:29.684845 4948 generic.go:334] "Generic (PLEG): container finished" podID="8a6ce79a-afdb-4e21-9e24-8de6924279bf" containerID="8157ef88958b3dff32dedd527eed80aa70a58f4f9fa4240a9b952e786b2b6880" exitCode=0 Mar 12 00:10:29 crc kubenswrapper[4948]: I0312 00:10:29.688233 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-btd4j" event={"ID":"8a6ce79a-afdb-4e21-9e24-8de6924279bf","Type":"ContainerDied","Data":"8157ef88958b3dff32dedd527eed80aa70a58f4f9fa4240a9b952e786b2b6880"} Mar 12 00:10:29 crc kubenswrapper[4948]: I0312 00:10:29.688282 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-btd4j" event={"ID":"8a6ce79a-afdb-4e21-9e24-8de6924279bf","Type":"ContainerStarted","Data":"1e8fe393396ecc17edae319ad76c0b9fe95e7a850a2db2f6841ecafa4869716d"} Mar 12 00:10:30 crc kubenswrapper[4948]: I0312 00:10:30.098485 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-ingress/router-default-5444994796-7vhkr" Mar 12 00:10:30 crc kubenswrapper[4948]: I0312 00:10:30.106127 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ingress/router-default-5444994796-7vhkr" Mar 12 00:10:31 crc kubenswrapper[4948]: I0312 00:10:31.543724 4948 ???:1] "http: TLS handshake error from 192.168.126.11:37936: no serving certificate available for the kubelet" Mar 12 00:10:33 crc kubenswrapper[4948]: I0312 00:10:33.242014 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-dns/dns-default-prqnt" Mar 12 00:10:37 crc kubenswrapper[4948]: I0312 00:10:37.762682 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/downloads-7954f5f757-xdh6w" Mar 12 00:10:37 crc kubenswrapper[4948]: I0312 00:10:37.767909 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-console/console-f9d7485db-qq8gg" Mar 12 00:10:37 crc kubenswrapper[4948]: I0312 00:10:37.770905 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-console/console-f9d7485db-qq8gg" Mar 12 00:10:38 crc kubenswrapper[4948]: I0312 00:10:38.740893 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6w6fq" event={"ID":"529630b8-cc23-44b7-bafd-6ae1313877d1","Type":"ContainerStarted","Data":"f035a5c066dcd602a7c0553774a03c5978745c83f0997e160378b6f3b4c716d0"} Mar 12 00:10:38 crc kubenswrapper[4948]: I0312 00:10:38.741882 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"02288ecc-afa1-48aa-9a2d-cc78303ede88","Type":"ContainerStarted","Data":"2046154f8b6ca6a025f01d7e1a0015386ef3503c4d31b4ee7e6e681a637be797"} Mar 12 00:10:41 crc kubenswrapper[4948]: I0312 00:10:41.297154 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7bb5cdccc6-pwdcd"] Mar 12 00:10:41 crc kubenswrapper[4948]: I0312 00:10:41.297711 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-7bb5cdccc6-pwdcd" podUID="cd346d37-ec14-49f0-8e88-f73930b8bab8" containerName="controller-manager" containerID="cri-o://e52800c551192345b03ce2fcafba9af613b213a83f507e64b193ef7ccd50b11e" gracePeriod=30 Mar 12 00:10:41 crc kubenswrapper[4948]: I0312 00:10:41.336743 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-778c687499-dcjm2"] Mar 12 00:10:41 crc kubenswrapper[4948]: I0312 00:10:41.342140 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-778c687499-dcjm2" podUID="8f87a342-cf29-45fe-acf3-2bfba03bea24" containerName="route-controller-manager" containerID="cri-o://778babe80ddb79b698b590dae2eae24873266a62a5bdc2c6418e23eaa451fea0" gracePeriod=30 Mar 12 00:10:42 crc kubenswrapper[4948]: I0312 00:10:42.765913 4948 generic.go:334] "Generic (PLEG): container finished" podID="8f87a342-cf29-45fe-acf3-2bfba03bea24" containerID="778babe80ddb79b698b590dae2eae24873266a62a5bdc2c6418e23eaa451fea0" exitCode=0 Mar 12 00:10:42 crc kubenswrapper[4948]: I0312 00:10:42.765995 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-778c687499-dcjm2" event={"ID":"8f87a342-cf29-45fe-acf3-2bfba03bea24","Type":"ContainerDied","Data":"778babe80ddb79b698b590dae2eae24873266a62a5bdc2c6418e23eaa451fea0"} Mar 12 00:10:42 crc kubenswrapper[4948]: I0312 00:10:42.767874 4948 generic.go:334] "Generic (PLEG): container finished" podID="cd346d37-ec14-49f0-8e88-f73930b8bab8" containerID="e52800c551192345b03ce2fcafba9af613b213a83f507e64b193ef7ccd50b11e" exitCode=0 Mar 12 00:10:42 crc kubenswrapper[4948]: I0312 00:10:42.767902 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7bb5cdccc6-pwdcd" event={"ID":"cd346d37-ec14-49f0-8e88-f73930b8bab8","Type":"ContainerDied","Data":"e52800c551192345b03ce2fcafba9af613b213a83f507e64b193ef7ccd50b11e"} Mar 12 00:10:44 crc kubenswrapper[4948]: I0312 00:10:44.784929 4948 patch_prober.go:28] interesting pod/machine-config-daemon-m4xwc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 00:10:44 crc kubenswrapper[4948]: I0312 00:10:44.785405 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" podUID="7e483c2b-08f0-4e92-8e4a-b7281f30af3e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 00:10:44 crc kubenswrapper[4948]: E0312 00:10:44.875272 4948 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/openshift4/ose-cli:latest" Mar 12 00:10:44 crc kubenswrapper[4948]: E0312 00:10:44.875430 4948 kuberuntime_manager.go:1274] "Unhandled Error" err=< Mar 12 00:10:44 crc kubenswrapper[4948]: container &Container{Name:oc,Image:registry.redhat.io/openshift4/ose-cli:latest,Command:[/bin/bash -c oc get csr -o go-template='{{range .items}}{{if not .status}}{{.metadata.name}}{{"\n"}}{{end}}{{end}}' | xargs --no-run-if-empty oc adm certificate approve Mar 12 00:10:44 crc kubenswrapper[4948]: ],Args:[],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:kube-api-access-6dvsl,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:nil,Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:File,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod auto-csr-approver-29554570-2b2hb_openshift-infra(2fb324d6-44d4-440d-ae60-a5356a3a057f): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled Mar 12 00:10:44 crc kubenswrapper[4948]: > logger="UnhandledError" Mar 12 00:10:44 crc kubenswrapper[4948]: E0312 00:10:44.877388 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-infra/auto-csr-approver-29554570-2b2hb" podUID="2fb324d6-44d4-440d-ae60-a5356a3a057f" Mar 12 00:10:45 crc kubenswrapper[4948]: I0312 00:10:45.358520 4948 patch_prober.go:28] interesting pod/controller-manager-7bb5cdccc6-pwdcd container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.46:8443/healthz\": dial tcp 10.217.0.46:8443: i/o timeout (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 12 00:10:45 crc kubenswrapper[4948]: I0312 00:10:45.358838 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-7bb5cdccc6-pwdcd" podUID="cd346d37-ec14-49f0-8e88-f73930b8bab8" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.46:8443/healthz\": dial tcp 10.217.0.46:8443: i/o timeout (Client.Timeout exceeded while awaiting headers)" Mar 12 00:10:45 crc kubenswrapper[4948]: E0312 00:10:45.783975 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"oc\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/openshift4/ose-cli:latest\\\"\"" pod="openshift-infra/auto-csr-approver-29554570-2b2hb" podUID="2fb324d6-44d4-440d-ae60-a5356a3a057f" Mar 12 00:10:45 crc kubenswrapper[4948]: I0312 00:10:45.940831 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:10:47 crc kubenswrapper[4948]: I0312 00:10:47.426528 4948 patch_prober.go:28] interesting pod/route-controller-manager-778c687499-dcjm2 container/route-controller-manager namespace/openshift-route-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.51:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 12 00:10:47 crc kubenswrapper[4948]: I0312 00:10:47.426955 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-route-controller-manager/route-controller-manager-778c687499-dcjm2" podUID="8f87a342-cf29-45fe-acf3-2bfba03bea24" containerName="route-controller-manager" probeResult="failure" output="Get \"https://10.217.0.51:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 12 00:10:49 crc kubenswrapper[4948]: E0312 00:10:49.282824 4948 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Mar 12 00:10:49 crc kubenswrapper[4948]: E0312 00:10:49.283007 4948 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-lbkcx,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-v8h88_openshift-marketplace(5d83d568-3968-4ea6-b548-511c951defa5): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 12 00:10:49 crc kubenswrapper[4948]: E0312 00:10:49.286128 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-v8h88" podUID="5d83d568-3968-4ea6-b548-511c951defa5" Mar 12 00:10:51 crc kubenswrapper[4948]: E0312 00:10:51.209706 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-v8h88" podUID="5d83d568-3968-4ea6-b548-511c951defa5" Mar 12 00:10:51 crc kubenswrapper[4948]: I0312 00:10:51.292092 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7bb5cdccc6-pwdcd" Mar 12 00:10:51 crc kubenswrapper[4948]: I0312 00:10:51.301281 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-778c687499-dcjm2" Mar 12 00:10:51 crc kubenswrapper[4948]: E0312 00:10:51.302397 4948 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Mar 12 00:10:51 crc kubenswrapper[4948]: E0312 00:10:51.302592 4948 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-pk6rv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-hcvm5_openshift-marketplace(239cfe46-3a3a-4624-9aa2-ea7112f28947): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 12 00:10:51 crc kubenswrapper[4948]: E0312 00:10:51.303876 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-hcvm5" podUID="239cfe46-3a3a-4624-9aa2-ea7112f28947" Mar 12 00:10:51 crc kubenswrapper[4948]: I0312 00:10:51.343494 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-74f85d4774-vqx4h"] Mar 12 00:10:51 crc kubenswrapper[4948]: E0312 00:10:51.343725 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="cd346d37-ec14-49f0-8e88-f73930b8bab8" containerName="controller-manager" Mar 12 00:10:51 crc kubenswrapper[4948]: I0312 00:10:51.343737 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="cd346d37-ec14-49f0-8e88-f73930b8bab8" containerName="controller-manager" Mar 12 00:10:51 crc kubenswrapper[4948]: E0312 00:10:51.343750 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8f87a342-cf29-45fe-acf3-2bfba03bea24" containerName="route-controller-manager" Mar 12 00:10:51 crc kubenswrapper[4948]: I0312 00:10:51.343757 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="8f87a342-cf29-45fe-acf3-2bfba03bea24" containerName="route-controller-manager" Mar 12 00:10:51 crc kubenswrapper[4948]: I0312 00:10:51.343898 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="8f87a342-cf29-45fe-acf3-2bfba03bea24" containerName="route-controller-manager" Mar 12 00:10:51 crc kubenswrapper[4948]: I0312 00:10:51.343912 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="cd346d37-ec14-49f0-8e88-f73930b8bab8" containerName="controller-manager" Mar 12 00:10:51 crc kubenswrapper[4948]: I0312 00:10:51.344381 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-74f85d4774-vqx4h"] Mar 12 00:10:51 crc kubenswrapper[4948]: I0312 00:10:51.344542 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-74f85d4774-vqx4h" Mar 12 00:10:51 crc kubenswrapper[4948]: I0312 00:10:51.383748 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-kgjb7\" (UniqueName: \"kubernetes.io/projected/8f87a342-cf29-45fe-acf3-2bfba03bea24-kube-api-access-kgjb7\") pod \"8f87a342-cf29-45fe-acf3-2bfba03bea24\" (UID: \"8f87a342-cf29-45fe-acf3-2bfba03bea24\") " Mar 12 00:10:51 crc kubenswrapper[4948]: I0312 00:10:51.384081 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8f87a342-cf29-45fe-acf3-2bfba03bea24-client-ca\") pod \"8f87a342-cf29-45fe-acf3-2bfba03bea24\" (UID: \"8f87a342-cf29-45fe-acf3-2bfba03bea24\") " Mar 12 00:10:51 crc kubenswrapper[4948]: I0312 00:10:51.384113 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd346d37-ec14-49f0-8e88-f73930b8bab8-config\") pod \"cd346d37-ec14-49f0-8e88-f73930b8bab8\" (UID: \"cd346d37-ec14-49f0-8e88-f73930b8bab8\") " Mar 12 00:10:51 crc kubenswrapper[4948]: I0312 00:10:51.384196 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cd346d37-ec14-49f0-8e88-f73930b8bab8-serving-cert\") pod \"cd346d37-ec14-49f0-8e88-f73930b8bab8\" (UID: \"cd346d37-ec14-49f0-8e88-f73930b8bab8\") " Mar 12 00:10:51 crc kubenswrapper[4948]: I0312 00:10:51.384226 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f87a342-cf29-45fe-acf3-2bfba03bea24-config\") pod \"8f87a342-cf29-45fe-acf3-2bfba03bea24\" (UID: \"8f87a342-cf29-45fe-acf3-2bfba03bea24\") " Mar 12 00:10:51 crc kubenswrapper[4948]: I0312 00:10:51.384255 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-mplgq\" (UniqueName: \"kubernetes.io/projected/cd346d37-ec14-49f0-8e88-f73930b8bab8-kube-api-access-mplgq\") pod \"cd346d37-ec14-49f0-8e88-f73930b8bab8\" (UID: \"cd346d37-ec14-49f0-8e88-f73930b8bab8\") " Mar 12 00:10:51 crc kubenswrapper[4948]: I0312 00:10:51.384316 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f87a342-cf29-45fe-acf3-2bfba03bea24-serving-cert\") pod \"8f87a342-cf29-45fe-acf3-2bfba03bea24\" (UID: \"8f87a342-cf29-45fe-acf3-2bfba03bea24\") " Mar 12 00:10:51 crc kubenswrapper[4948]: I0312 00:10:51.384354 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/cd346d37-ec14-49f0-8e88-f73930b8bab8-proxy-ca-bundles\") pod \"cd346d37-ec14-49f0-8e88-f73930b8bab8\" (UID: \"cd346d37-ec14-49f0-8e88-f73930b8bab8\") " Mar 12 00:10:51 crc kubenswrapper[4948]: I0312 00:10:51.384386 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cd346d37-ec14-49f0-8e88-f73930b8bab8-client-ca\") pod \"cd346d37-ec14-49f0-8e88-f73930b8bab8\" (UID: \"cd346d37-ec14-49f0-8e88-f73930b8bab8\") " Mar 12 00:10:51 crc kubenswrapper[4948]: I0312 00:10:51.384577 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-64qjf\" (UniqueName: \"kubernetes.io/projected/6ddbb31b-e927-4588-94f2-c2ffc5b4e604-kube-api-access-64qjf\") pod \"controller-manager-74f85d4774-vqx4h\" (UID: \"6ddbb31b-e927-4588-94f2-c2ffc5b4e604\") " pod="openshift-controller-manager/controller-manager-74f85d4774-vqx4h" Mar 12 00:10:51 crc kubenswrapper[4948]: I0312 00:10:51.384605 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6ddbb31b-e927-4588-94f2-c2ffc5b4e604-proxy-ca-bundles\") pod \"controller-manager-74f85d4774-vqx4h\" (UID: \"6ddbb31b-e927-4588-94f2-c2ffc5b4e604\") " pod="openshift-controller-manager/controller-manager-74f85d4774-vqx4h" Mar 12 00:10:51 crc kubenswrapper[4948]: I0312 00:10:51.384647 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ddbb31b-e927-4588-94f2-c2ffc5b4e604-config\") pod \"controller-manager-74f85d4774-vqx4h\" (UID: \"6ddbb31b-e927-4588-94f2-c2ffc5b4e604\") " pod="openshift-controller-manager/controller-manager-74f85d4774-vqx4h" Mar 12 00:10:51 crc kubenswrapper[4948]: I0312 00:10:51.384672 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6ddbb31b-e927-4588-94f2-c2ffc5b4e604-client-ca\") pod \"controller-manager-74f85d4774-vqx4h\" (UID: \"6ddbb31b-e927-4588-94f2-c2ffc5b4e604\") " pod="openshift-controller-manager/controller-manager-74f85d4774-vqx4h" Mar 12 00:10:51 crc kubenswrapper[4948]: I0312 00:10:51.384692 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6ddbb31b-e927-4588-94f2-c2ffc5b4e604-serving-cert\") pod \"controller-manager-74f85d4774-vqx4h\" (UID: \"6ddbb31b-e927-4588-94f2-c2ffc5b4e604\") " pod="openshift-controller-manager/controller-manager-74f85d4774-vqx4h" Mar 12 00:10:51 crc kubenswrapper[4948]: I0312 00:10:51.385275 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f87a342-cf29-45fe-acf3-2bfba03bea24-client-ca" (OuterVolumeSpecName: "client-ca") pod "8f87a342-cf29-45fe-acf3-2bfba03bea24" (UID: "8f87a342-cf29-45fe-acf3-2bfba03bea24"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:10:51 crc kubenswrapper[4948]: I0312 00:10:51.385989 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd346d37-ec14-49f0-8e88-f73930b8bab8-client-ca" (OuterVolumeSpecName: "client-ca") pod "cd346d37-ec14-49f0-8e88-f73930b8bab8" (UID: "cd346d37-ec14-49f0-8e88-f73930b8bab8"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:10:51 crc kubenswrapper[4948]: I0312 00:10:51.386054 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/8f87a342-cf29-45fe-acf3-2bfba03bea24-config" (OuterVolumeSpecName: "config") pod "8f87a342-cf29-45fe-acf3-2bfba03bea24" (UID: "8f87a342-cf29-45fe-acf3-2bfba03bea24"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:10:51 crc kubenswrapper[4948]: I0312 00:10:51.386670 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd346d37-ec14-49f0-8e88-f73930b8bab8-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "cd346d37-ec14-49f0-8e88-f73930b8bab8" (UID: "cd346d37-ec14-49f0-8e88-f73930b8bab8"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:10:51 crc kubenswrapper[4948]: I0312 00:10:51.386683 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/cd346d37-ec14-49f0-8e88-f73930b8bab8-config" (OuterVolumeSpecName: "config") pod "cd346d37-ec14-49f0-8e88-f73930b8bab8" (UID: "cd346d37-ec14-49f0-8e88-f73930b8bab8"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:10:51 crc kubenswrapper[4948]: I0312 00:10:51.391096 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/cd346d37-ec14-49f0-8e88-f73930b8bab8-kube-api-access-mplgq" (OuterVolumeSpecName: "kube-api-access-mplgq") pod "cd346d37-ec14-49f0-8e88-f73930b8bab8" (UID: "cd346d37-ec14-49f0-8e88-f73930b8bab8"). InnerVolumeSpecName "kube-api-access-mplgq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:10:51 crc kubenswrapper[4948]: I0312 00:10:51.392033 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8f87a342-cf29-45fe-acf3-2bfba03bea24-kube-api-access-kgjb7" (OuterVolumeSpecName: "kube-api-access-kgjb7") pod "8f87a342-cf29-45fe-acf3-2bfba03bea24" (UID: "8f87a342-cf29-45fe-acf3-2bfba03bea24"). InnerVolumeSpecName "kube-api-access-kgjb7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:10:51 crc kubenswrapper[4948]: I0312 00:10:51.394219 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/cd346d37-ec14-49f0-8e88-f73930b8bab8-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "cd346d37-ec14-49f0-8e88-f73930b8bab8" (UID: "cd346d37-ec14-49f0-8e88-f73930b8bab8"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:10:51 crc kubenswrapper[4948]: I0312 00:10:51.395782 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/8f87a342-cf29-45fe-acf3-2bfba03bea24-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "8f87a342-cf29-45fe-acf3-2bfba03bea24" (UID: "8f87a342-cf29-45fe-acf3-2bfba03bea24"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:10:51 crc kubenswrapper[4948]: E0312 00:10:51.400622 4948 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/community-operator-index:v4.18" Mar 12 00:10:51 crc kubenswrapper[4948]: E0312 00:10:51.400762 4948 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/community-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-5gmsv,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod community-operators-62wnf_openshift-marketplace(e6318b61-ef2d-41da-8c92-ff7523ce613f): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 12 00:10:51 crc kubenswrapper[4948]: E0312 00:10:51.401913 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/community-operators-62wnf" podUID="e6318b61-ef2d-41da-8c92-ff7523ce613f" Mar 12 00:10:51 crc kubenswrapper[4948]: E0312 00:10:51.402280 4948 log.go:32] "PullImage from image service failed" err="rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" image="registry.redhat.io/redhat/certified-operator-index:v4.18" Mar 12 00:10:51 crc kubenswrapper[4948]: E0312 00:10:51.402641 4948 kuberuntime_manager.go:1274] "Unhandled Error" err="init container &Container{Name:extract-content,Image:registry.redhat.io/redhat/certified-operator-index:v4.18,Command:[/utilities/copy-content],Args:[--catalog.from=/configs --catalog.to=/extracted-catalog/catalog --cache.from=/tmp/cache --cache.to=/extracted-catalog/cache],WorkingDir:,Ports:[]ContainerPort{},Env:[]EnvVar{},Resources:ResourceRequirements{Limits:ResourceList{},Requests:ResourceList{},Claims:[]ResourceClaim{},},VolumeMounts:[]VolumeMount{VolumeMount{Name:utilities,ReadOnly:false,MountPath:/utilities,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:catalog-content,ReadOnly:false,MountPath:/extracted-catalog,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},VolumeMount{Name:kube-api-access-rljd9,ReadOnly:true,MountPath:/var/run/secrets/kubernetes.io/serviceaccount,SubPath:,MountPropagation:nil,SubPathExpr:,RecursiveReadOnly:nil,},},LivenessProbe:nil,ReadinessProbe:nil,Lifecycle:nil,TerminationMessagePath:/dev/termination-log,ImagePullPolicy:Always,SecurityContext:&SecurityContext{Capabilities:&Capabilities{Add:[],Drop:[ALL],},Privileged:nil,SELinuxOptions:nil,RunAsUser:*1000170000,RunAsNonRoot:*true,ReadOnlyRootFilesystem:nil,AllowPrivilegeEscalation:*false,RunAsGroup:nil,ProcMount:nil,WindowsOptions:nil,SeccompProfile:nil,AppArmorProfile:nil,},Stdin:false,StdinOnce:false,TTY:false,EnvFrom:[]EnvFromSource{},TerminationMessagePolicy:FallbackToLogsOnError,VolumeDevices:[]VolumeDevice{},StartupProbe:nil,ResizePolicy:[]ContainerResizePolicy{},RestartPolicy:nil,} start failed in pod certified-operators-rflhd_openshift-marketplace(5184125a-dde1-4db4-9d6b-5b3c45d8d82a): ErrImagePull: rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled" logger="UnhandledError" Mar 12 00:10:51 crc kubenswrapper[4948]: E0312 00:10:51.403801 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ErrImagePull: \"rpc error: code = Canceled desc = copying system image from manifest list: copying config: context canceled\"" pod="openshift-marketplace/certified-operators-rflhd" podUID="5184125a-dde1-4db4-9d6b-5b3c45d8d82a" Mar 12 00:10:51 crc kubenswrapper[4948]: I0312 00:10:51.424665 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-controller-manager/revision-pruner-9-crc"] Mar 12 00:10:51 crc kubenswrapper[4948]: I0312 00:10:51.487593 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6ddbb31b-e927-4588-94f2-c2ffc5b4e604-serving-cert\") pod \"controller-manager-74f85d4774-vqx4h\" (UID: \"6ddbb31b-e927-4588-94f2-c2ffc5b4e604\") " pod="openshift-controller-manager/controller-manager-74f85d4774-vqx4h" Mar 12 00:10:51 crc kubenswrapper[4948]: I0312 00:10:51.487680 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-64qjf\" (UniqueName: \"kubernetes.io/projected/6ddbb31b-e927-4588-94f2-c2ffc5b4e604-kube-api-access-64qjf\") pod \"controller-manager-74f85d4774-vqx4h\" (UID: \"6ddbb31b-e927-4588-94f2-c2ffc5b4e604\") " pod="openshift-controller-manager/controller-manager-74f85d4774-vqx4h" Mar 12 00:10:51 crc kubenswrapper[4948]: I0312 00:10:51.487705 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6ddbb31b-e927-4588-94f2-c2ffc5b4e604-proxy-ca-bundles\") pod \"controller-manager-74f85d4774-vqx4h\" (UID: \"6ddbb31b-e927-4588-94f2-c2ffc5b4e604\") " pod="openshift-controller-manager/controller-manager-74f85d4774-vqx4h" Mar 12 00:10:51 crc kubenswrapper[4948]: I0312 00:10:51.487735 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ddbb31b-e927-4588-94f2-c2ffc5b4e604-config\") pod \"controller-manager-74f85d4774-vqx4h\" (UID: \"6ddbb31b-e927-4588-94f2-c2ffc5b4e604\") " pod="openshift-controller-manager/controller-manager-74f85d4774-vqx4h" Mar 12 00:10:51 crc kubenswrapper[4948]: I0312 00:10:51.487760 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6ddbb31b-e927-4588-94f2-c2ffc5b4e604-client-ca\") pod \"controller-manager-74f85d4774-vqx4h\" (UID: \"6ddbb31b-e927-4588-94f2-c2ffc5b4e604\") " pod="openshift-controller-manager/controller-manager-74f85d4774-vqx4h" Mar 12 00:10:51 crc kubenswrapper[4948]: I0312 00:10:51.487794 4948 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/8f87a342-cf29-45fe-acf3-2bfba03bea24-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 00:10:51 crc kubenswrapper[4948]: I0312 00:10:51.487807 4948 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/cd346d37-ec14-49f0-8e88-f73930b8bab8-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 12 00:10:51 crc kubenswrapper[4948]: I0312 00:10:51.487816 4948 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/cd346d37-ec14-49f0-8e88-f73930b8bab8-client-ca\") on node \"crc\" DevicePath \"\"" Mar 12 00:10:51 crc kubenswrapper[4948]: I0312 00:10:51.487825 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-kgjb7\" (UniqueName: \"kubernetes.io/projected/8f87a342-cf29-45fe-acf3-2bfba03bea24-kube-api-access-kgjb7\") on node \"crc\" DevicePath \"\"" Mar 12 00:10:51 crc kubenswrapper[4948]: I0312 00:10:51.487833 4948 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/8f87a342-cf29-45fe-acf3-2bfba03bea24-client-ca\") on node \"crc\" DevicePath \"\"" Mar 12 00:10:51 crc kubenswrapper[4948]: I0312 00:10:51.487843 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/cd346d37-ec14-49f0-8e88-f73930b8bab8-config\") on node \"crc\" DevicePath \"\"" Mar 12 00:10:51 crc kubenswrapper[4948]: I0312 00:10:51.487852 4948 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/cd346d37-ec14-49f0-8e88-f73930b8bab8-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 00:10:51 crc kubenswrapper[4948]: I0312 00:10:51.487860 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/8f87a342-cf29-45fe-acf3-2bfba03bea24-config\") on node \"crc\" DevicePath \"\"" Mar 12 00:10:51 crc kubenswrapper[4948]: I0312 00:10:51.487870 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-mplgq\" (UniqueName: \"kubernetes.io/projected/cd346d37-ec14-49f0-8e88-f73930b8bab8-kube-api-access-mplgq\") on node \"crc\" DevicePath \"\"" Mar 12 00:10:51 crc kubenswrapper[4948]: I0312 00:10:51.489215 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ddbb31b-e927-4588-94f2-c2ffc5b4e604-config\") pod \"controller-manager-74f85d4774-vqx4h\" (UID: \"6ddbb31b-e927-4588-94f2-c2ffc5b4e604\") " pod="openshift-controller-manager/controller-manager-74f85d4774-vqx4h" Mar 12 00:10:51 crc kubenswrapper[4948]: I0312 00:10:51.489968 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6ddbb31b-e927-4588-94f2-c2ffc5b4e604-client-ca\") pod \"controller-manager-74f85d4774-vqx4h\" (UID: \"6ddbb31b-e927-4588-94f2-c2ffc5b4e604\") " pod="openshift-controller-manager/controller-manager-74f85d4774-vqx4h" Mar 12 00:10:51 crc kubenswrapper[4948]: I0312 00:10:51.490285 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6ddbb31b-e927-4588-94f2-c2ffc5b4e604-proxy-ca-bundles\") pod \"controller-manager-74f85d4774-vqx4h\" (UID: \"6ddbb31b-e927-4588-94f2-c2ffc5b4e604\") " pod="openshift-controller-manager/controller-manager-74f85d4774-vqx4h" Mar 12 00:10:51 crc kubenswrapper[4948]: I0312 00:10:51.492331 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6ddbb31b-e927-4588-94f2-c2ffc5b4e604-serving-cert\") pod \"controller-manager-74f85d4774-vqx4h\" (UID: \"6ddbb31b-e927-4588-94f2-c2ffc5b4e604\") " pod="openshift-controller-manager/controller-manager-74f85d4774-vqx4h" Mar 12 00:10:51 crc kubenswrapper[4948]: I0312 00:10:51.501626 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-64qjf\" (UniqueName: \"kubernetes.io/projected/6ddbb31b-e927-4588-94f2-c2ffc5b4e604-kube-api-access-64qjf\") pod \"controller-manager-74f85d4774-vqx4h\" (UID: \"6ddbb31b-e927-4588-94f2-c2ffc5b4e604\") " pod="openshift-controller-manager/controller-manager-74f85d4774-vqx4h" Mar 12 00:10:51 crc kubenswrapper[4948]: I0312 00:10:51.657712 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-74f85d4774-vqx4h" Mar 12 00:10:51 crc kubenswrapper[4948]: I0312 00:10:51.816513 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-778c687499-dcjm2" event={"ID":"8f87a342-cf29-45fe-acf3-2bfba03bea24","Type":"ContainerDied","Data":"b9b1a41b07b3648b422714ea2b8280dcdbf2062fc282b54a5bf8486d324cc6e2"} Mar 12 00:10:51 crc kubenswrapper[4948]: I0312 00:10:51.816536 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-778c687499-dcjm2" Mar 12 00:10:51 crc kubenswrapper[4948]: I0312 00:10:51.816566 4948 scope.go:117] "RemoveContainer" containerID="778babe80ddb79b698b590dae2eae24873266a62a5bdc2c6418e23eaa451fea0" Mar 12 00:10:51 crc kubenswrapper[4948]: I0312 00:10:51.818511 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7bb5cdccc6-pwdcd" event={"ID":"cd346d37-ec14-49f0-8e88-f73930b8bab8","Type":"ContainerDied","Data":"4ef442baa04f491a58e34e7e726e281ce5c5a6d127017aca94eaf63dceefc1f0"} Mar 12 00:10:51 crc kubenswrapper[4948]: I0312 00:10:51.818568 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7bb5cdccc6-pwdcd" Mar 12 00:10:51 crc kubenswrapper[4948]: I0312 00:10:51.820994 4948 generic.go:334] "Generic (PLEG): container finished" podID="2c8b7f32-4f80-4f95-b361-7d4687416711" containerID="6c8849999918cbc4ac96da6f6d31b0a3594724331620293f8feb102cd4fadaf5" exitCode=0 Mar 12 00:10:51 crc kubenswrapper[4948]: I0312 00:10:51.821026 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29554560-slpk9" event={"ID":"2c8b7f32-4f80-4f95-b361-7d4687416711","Type":"ContainerDied","Data":"6c8849999918cbc4ac96da6f6d31b0a3594724331620293f8feb102cd4fadaf5"} Mar 12 00:10:51 crc kubenswrapper[4948]: I0312 00:10:51.912225 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7bb5cdccc6-pwdcd"] Mar 12 00:10:51 crc kubenswrapper[4948]: I0312 00:10:51.917948 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-7bb5cdccc6-pwdcd"] Mar 12 00:10:51 crc kubenswrapper[4948]: I0312 00:10:51.921480 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-778c687499-dcjm2"] Mar 12 00:10:51 crc kubenswrapper[4948]: I0312 00:10:51.923936 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-778c687499-dcjm2"] Mar 12 00:10:52 crc kubenswrapper[4948]: I0312 00:10:52.042519 4948 ???:1] "http: TLS handshake error from 192.168.126.11:34896: no serving certificate available for the kubelet" Mar 12 00:10:53 crc kubenswrapper[4948]: E0312 00:10:53.109854 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-rflhd" podUID="5184125a-dde1-4db4-9d6b-5b3c45d8d82a" Mar 12 00:10:53 crc kubenswrapper[4948]: E0312 00:10:53.109923 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/community-operator-index:v4.18\\\"\"" pod="openshift-marketplace/community-operators-62wnf" podUID="e6318b61-ef2d-41da-8c92-ff7523ce613f" Mar 12 00:10:53 crc kubenswrapper[4948]: E0312 00:10:53.119968 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"extract-content\" with ImagePullBackOff: \"Back-off pulling image \\\"registry.redhat.io/redhat/certified-operator-index:v4.18\\\"\"" pod="openshift-marketplace/certified-operators-hcvm5" podUID="239cfe46-3a3a-4624-9aa2-ea7112f28947" Mar 12 00:10:53 crc kubenswrapper[4948]: W0312 00:10:53.135100 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-pod44d3c29e_5c1e_41be_ae1a_8d1e5f9e39c7.slice/crio-5e78897f10b9b3c3672515acd885ca4d0edda2efc255ea97ead64bc5ba2f6aca WatchSource:0}: Error finding container 5e78897f10b9b3c3672515acd885ca4d0edda2efc255ea97ead64bc5ba2f6aca: Status 404 returned error can't find the container with id 5e78897f10b9b3c3672515acd885ca4d0edda2efc255ea97ead64bc5ba2f6aca Mar 12 00:10:53 crc kubenswrapper[4948]: I0312 00:10:53.178856 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29554560-slpk9" Mar 12 00:10:53 crc kubenswrapper[4948]: I0312 00:10:53.234027 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lbrg4\" (UniqueName: \"kubernetes.io/projected/2c8b7f32-4f80-4f95-b361-7d4687416711-kube-api-access-lbrg4\") pod \"2c8b7f32-4f80-4f95-b361-7d4687416711\" (UID: \"2c8b7f32-4f80-4f95-b361-7d4687416711\") " Mar 12 00:10:53 crc kubenswrapper[4948]: I0312 00:10:53.234138 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/2c8b7f32-4f80-4f95-b361-7d4687416711-serviceca\") pod \"2c8b7f32-4f80-4f95-b361-7d4687416711\" (UID: \"2c8b7f32-4f80-4f95-b361-7d4687416711\") " Mar 12 00:10:53 crc kubenswrapper[4948]: I0312 00:10:53.235299 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/2c8b7f32-4f80-4f95-b361-7d4687416711-serviceca" (OuterVolumeSpecName: "serviceca") pod "2c8b7f32-4f80-4f95-b361-7d4687416711" (UID: "2c8b7f32-4f80-4f95-b361-7d4687416711"). InnerVolumeSpecName "serviceca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:10:53 crc kubenswrapper[4948]: I0312 00:10:53.262514 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2c8b7f32-4f80-4f95-b361-7d4687416711-kube-api-access-lbrg4" (OuterVolumeSpecName: "kube-api-access-lbrg4") pod "2c8b7f32-4f80-4f95-b361-7d4687416711" (UID: "2c8b7f32-4f80-4f95-b361-7d4687416711"). InnerVolumeSpecName "kube-api-access-lbrg4". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:10:53 crc kubenswrapper[4948]: I0312 00:10:53.327957 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8f87a342-cf29-45fe-acf3-2bfba03bea24" path="/var/lib/kubelet/pods/8f87a342-cf29-45fe-acf3-2bfba03bea24/volumes" Mar 12 00:10:53 crc kubenswrapper[4948]: I0312 00:10:53.328810 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="cd346d37-ec14-49f0-8e88-f73930b8bab8" path="/var/lib/kubelet/pods/cd346d37-ec14-49f0-8e88-f73930b8bab8/volumes" Mar 12 00:10:53 crc kubenswrapper[4948]: I0312 00:10:53.335413 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lbrg4\" (UniqueName: \"kubernetes.io/projected/2c8b7f32-4f80-4f95-b361-7d4687416711-kube-api-access-lbrg4\") on node \"crc\" DevicePath \"\"" Mar 12 00:10:53 crc kubenswrapper[4948]: I0312 00:10:53.335457 4948 reconciler_common.go:293] "Volume detached for volume \"serviceca\" (UniqueName: \"kubernetes.io/configmap/2c8b7f32-4f80-4f95-b361-7d4687416711-serviceca\") on node \"crc\" DevicePath \"\"" Mar 12 00:10:53 crc kubenswrapper[4948]: I0312 00:10:53.830409 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"44d3c29e-5c1e-41be-ae1a-8d1e5f9e39c7","Type":"ContainerStarted","Data":"5e78897f10b9b3c3672515acd885ca4d0edda2efc255ea97ead64bc5ba2f6aca"} Mar 12 00:10:53 crc kubenswrapper[4948]: I0312 00:10:53.832177 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-pruner-29554560-slpk9" event={"ID":"2c8b7f32-4f80-4f95-b361-7d4687416711","Type":"ContainerDied","Data":"fc302ca7ca1a53ca39b77dc45506f5093ecef4f025d778c8e2a56f7563e9a34d"} Mar 12 00:10:53 crc kubenswrapper[4948]: I0312 00:10:53.832201 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fc302ca7ca1a53ca39b77dc45506f5093ecef4f025d778c8e2a56f7563e9a34d" Mar 12 00:10:53 crc kubenswrapper[4948]: I0312 00:10:53.832245 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-pruner-29554560-slpk9" Mar 12 00:10:54 crc kubenswrapper[4948]: I0312 00:10:54.245697 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5488b469ff-xfnxl"] Mar 12 00:10:54 crc kubenswrapper[4948]: E0312 00:10:54.245893 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2c8b7f32-4f80-4f95-b361-7d4687416711" containerName="image-pruner" Mar 12 00:10:54 crc kubenswrapper[4948]: I0312 00:10:54.245905 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="2c8b7f32-4f80-4f95-b361-7d4687416711" containerName="image-pruner" Mar 12 00:10:54 crc kubenswrapper[4948]: I0312 00:10:54.245998 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="2c8b7f32-4f80-4f95-b361-7d4687416711" containerName="image-pruner" Mar 12 00:10:54 crc kubenswrapper[4948]: I0312 00:10:54.246390 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5488b469ff-xfnxl" Mar 12 00:10:54 crc kubenswrapper[4948]: I0312 00:10:54.249736 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 12 00:10:54 crc kubenswrapper[4948]: I0312 00:10:54.250180 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 12 00:10:54 crc kubenswrapper[4948]: I0312 00:10:54.263502 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 12 00:10:54 crc kubenswrapper[4948]: I0312 00:10:54.263621 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 12 00:10:54 crc kubenswrapper[4948]: I0312 00:10:54.263498 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 12 00:10:54 crc kubenswrapper[4948]: I0312 00:10:54.263901 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 12 00:10:54 crc kubenswrapper[4948]: I0312 00:10:54.276290 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5488b469ff-xfnxl"] Mar 12 00:10:54 crc kubenswrapper[4948]: I0312 00:10:54.348705 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8j7s6\" (UniqueName: \"kubernetes.io/projected/850a971c-3fcf-428e-8e1a-2a3a22e4afa2-kube-api-access-8j7s6\") pod \"route-controller-manager-5488b469ff-xfnxl\" (UID: \"850a971c-3fcf-428e-8e1a-2a3a22e4afa2\") " pod="openshift-route-controller-manager/route-controller-manager-5488b469ff-xfnxl" Mar 12 00:10:54 crc kubenswrapper[4948]: I0312 00:10:54.348787 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/850a971c-3fcf-428e-8e1a-2a3a22e4afa2-client-ca\") pod \"route-controller-manager-5488b469ff-xfnxl\" (UID: \"850a971c-3fcf-428e-8e1a-2a3a22e4afa2\") " pod="openshift-route-controller-manager/route-controller-manager-5488b469ff-xfnxl" Mar 12 00:10:54 crc kubenswrapper[4948]: I0312 00:10:54.348908 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/850a971c-3fcf-428e-8e1a-2a3a22e4afa2-config\") pod \"route-controller-manager-5488b469ff-xfnxl\" (UID: \"850a971c-3fcf-428e-8e1a-2a3a22e4afa2\") " pod="openshift-route-controller-manager/route-controller-manager-5488b469ff-xfnxl" Mar 12 00:10:54 crc kubenswrapper[4948]: I0312 00:10:54.349018 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/850a971c-3fcf-428e-8e1a-2a3a22e4afa2-serving-cert\") pod \"route-controller-manager-5488b469ff-xfnxl\" (UID: \"850a971c-3fcf-428e-8e1a-2a3a22e4afa2\") " pod="openshift-route-controller-manager/route-controller-manager-5488b469ff-xfnxl" Mar 12 00:10:54 crc kubenswrapper[4948]: I0312 00:10:54.450681 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8j7s6\" (UniqueName: \"kubernetes.io/projected/850a971c-3fcf-428e-8e1a-2a3a22e4afa2-kube-api-access-8j7s6\") pod \"route-controller-manager-5488b469ff-xfnxl\" (UID: \"850a971c-3fcf-428e-8e1a-2a3a22e4afa2\") " pod="openshift-route-controller-manager/route-controller-manager-5488b469ff-xfnxl" Mar 12 00:10:54 crc kubenswrapper[4948]: I0312 00:10:54.450961 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/850a971c-3fcf-428e-8e1a-2a3a22e4afa2-client-ca\") pod \"route-controller-manager-5488b469ff-xfnxl\" (UID: \"850a971c-3fcf-428e-8e1a-2a3a22e4afa2\") " pod="openshift-route-controller-manager/route-controller-manager-5488b469ff-xfnxl" Mar 12 00:10:54 crc kubenswrapper[4948]: I0312 00:10:54.450987 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/850a971c-3fcf-428e-8e1a-2a3a22e4afa2-config\") pod \"route-controller-manager-5488b469ff-xfnxl\" (UID: \"850a971c-3fcf-428e-8e1a-2a3a22e4afa2\") " pod="openshift-route-controller-manager/route-controller-manager-5488b469ff-xfnxl" Mar 12 00:10:54 crc kubenswrapper[4948]: I0312 00:10:54.451008 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/850a971c-3fcf-428e-8e1a-2a3a22e4afa2-serving-cert\") pod \"route-controller-manager-5488b469ff-xfnxl\" (UID: \"850a971c-3fcf-428e-8e1a-2a3a22e4afa2\") " pod="openshift-route-controller-manager/route-controller-manager-5488b469ff-xfnxl" Mar 12 00:10:54 crc kubenswrapper[4948]: I0312 00:10:54.452009 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/850a971c-3fcf-428e-8e1a-2a3a22e4afa2-client-ca\") pod \"route-controller-manager-5488b469ff-xfnxl\" (UID: \"850a971c-3fcf-428e-8e1a-2a3a22e4afa2\") " pod="openshift-route-controller-manager/route-controller-manager-5488b469ff-xfnxl" Mar 12 00:10:54 crc kubenswrapper[4948]: I0312 00:10:54.452090 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/850a971c-3fcf-428e-8e1a-2a3a22e4afa2-config\") pod \"route-controller-manager-5488b469ff-xfnxl\" (UID: \"850a971c-3fcf-428e-8e1a-2a3a22e4afa2\") " pod="openshift-route-controller-manager/route-controller-manager-5488b469ff-xfnxl" Mar 12 00:10:54 crc kubenswrapper[4948]: I0312 00:10:54.457199 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/850a971c-3fcf-428e-8e1a-2a3a22e4afa2-serving-cert\") pod \"route-controller-manager-5488b469ff-xfnxl\" (UID: \"850a971c-3fcf-428e-8e1a-2a3a22e4afa2\") " pod="openshift-route-controller-manager/route-controller-manager-5488b469ff-xfnxl" Mar 12 00:10:54 crc kubenswrapper[4948]: I0312 00:10:54.467039 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8j7s6\" (UniqueName: \"kubernetes.io/projected/850a971c-3fcf-428e-8e1a-2a3a22e4afa2-kube-api-access-8j7s6\") pod \"route-controller-manager-5488b469ff-xfnxl\" (UID: \"850a971c-3fcf-428e-8e1a-2a3a22e4afa2\") " pod="openshift-route-controller-manager/route-controller-manager-5488b469ff-xfnxl" Mar 12 00:10:54 crc kubenswrapper[4948]: I0312 00:10:54.582869 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5488b469ff-xfnxl" Mar 12 00:10:56 crc kubenswrapper[4948]: I0312 00:10:56.370090 4948 scope.go:117] "RemoveContainer" containerID="e52800c551192345b03ce2fcafba9af613b213a83f507e64b193ef7ccd50b11e" Mar 12 00:10:56 crc kubenswrapper[4948]: I0312 00:10:56.844921 4948 generic.go:334] "Generic (PLEG): container finished" podID="529630b8-cc23-44b7-bafd-6ae1313877d1" containerID="e18323d8a5d040179826f150b6099523db491ee1abb8b59ec1cb4aa4f795efdd" exitCode=0 Mar 12 00:10:56 crc kubenswrapper[4948]: I0312 00:10:56.845135 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6w6fq" event={"ID":"529630b8-cc23-44b7-bafd-6ae1313877d1","Type":"ContainerDied","Data":"e18323d8a5d040179826f150b6099523db491ee1abb8b59ec1cb4aa4f795efdd"} Mar 12 00:10:56 crc kubenswrapper[4948]: I0312 00:10:56.846768 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"44d3c29e-5c1e-41be-ae1a-8d1e5f9e39c7","Type":"ContainerStarted","Data":"0a0c38f4d0f9ca9157026aed00d883062d2184f3d7694889696a5eae1fcc82b2"} Mar 12 00:10:56 crc kubenswrapper[4948]: I0312 00:10:56.850863 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bsmcp" event={"ID":"44212e8e-c675-4643-a752-997be5f8e980","Type":"ContainerStarted","Data":"b87f1313ce649ed1dd52e80afc20736b50c50e9bf8e2e84c9837a9177cd604ba"} Mar 12 00:10:56 crc kubenswrapper[4948]: I0312 00:10:56.854252 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zc9zh" event={"ID":"f8cf7ee1-e662-4afd-9e08-112171677c22","Type":"ContainerStarted","Data":"53375f01d7cb1d105553e223936295bb8066f47dbacf62b87c6fd8cc659f2202"} Mar 12 00:10:56 crc kubenswrapper[4948]: I0312 00:10:56.856238 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-btd4j" event={"ID":"8a6ce79a-afdb-4e21-9e24-8de6924279bf","Type":"ContainerStarted","Data":"f671d445fb2b1163a7bd29f7448a3d3c3956cda4379bfe37ce56af3269fe0c79"} Mar 12 00:10:56 crc kubenswrapper[4948]: I0312 00:10:56.859515 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"02288ecc-afa1-48aa-9a2d-cc78303ede88","Type":"ContainerStarted","Data":"805fd24d2606a526eaa0631c6bb97d4b0bc682d491f077df0e5adc45ab3b8b25"} Mar 12 00:10:56 crc kubenswrapper[4948]: I0312 00:10:56.891041 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/revision-pruner-8-crc" podStartSLOduration=28.891021735 podStartE2EDuration="28.891021735s" podCreationTimestamp="2026-03-12 00:10:28 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:10:56.888309342 +0000 UTC m=+256.343913080" watchObservedRunningTime="2026-03-12 00:10:56.891021735 +0000 UTC m=+256.346625473" Mar 12 00:10:56 crc kubenswrapper[4948]: W0312 00:10:56.900973 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod6ddbb31b_e927_4588_94f2_c2ffc5b4e604.slice/crio-863403b3b5a934359a7de9563e1ba0d16821bf9ffc8a5e3a2907303c7758992c WatchSource:0}: Error finding container 863403b3b5a934359a7de9563e1ba0d16821bf9ffc8a5e3a2907303c7758992c: Status 404 returned error can't find the container with id 863403b3b5a934359a7de9563e1ba0d16821bf9ffc8a5e3a2907303c7758992c Mar 12 00:10:56 crc kubenswrapper[4948]: I0312 00:10:56.902133 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-74f85d4774-vqx4h"] Mar 12 00:10:56 crc kubenswrapper[4948]: I0312 00:10:56.917223 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5488b469ff-xfnxl"] Mar 12 00:10:56 crc kubenswrapper[4948]: W0312 00:10:56.929428 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod850a971c_3fcf_428e_8e1a_2a3a22e4afa2.slice/crio-932f3f1fa8746f8a5ab5b1d0879e755ac7be65c516e805d635c85523a809a96c WatchSource:0}: Error finding container 932f3f1fa8746f8a5ab5b1d0879e755ac7be65c516e805d635c85523a809a96c: Status 404 returned error can't find the container with id 932f3f1fa8746f8a5ab5b1d0879e755ac7be65c516e805d635c85523a809a96c Mar 12 00:10:57 crc kubenswrapper[4948]: I0312 00:10:57.866625 4948 generic.go:334] "Generic (PLEG): container finished" podID="44212e8e-c675-4643-a752-997be5f8e980" containerID="b87f1313ce649ed1dd52e80afc20736b50c50e9bf8e2e84c9837a9177cd604ba" exitCode=0 Mar 12 00:10:57 crc kubenswrapper[4948]: I0312 00:10:57.866667 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bsmcp" event={"ID":"44212e8e-c675-4643-a752-997be5f8e980","Type":"ContainerDied","Data":"b87f1313ce649ed1dd52e80afc20736b50c50e9bf8e2e84c9837a9177cd604ba"} Mar 12 00:10:57 crc kubenswrapper[4948]: I0312 00:10:57.871323 4948 generic.go:334] "Generic (PLEG): container finished" podID="02288ecc-afa1-48aa-9a2d-cc78303ede88" containerID="805fd24d2606a526eaa0631c6bb97d4b0bc682d491f077df0e5adc45ab3b8b25" exitCode=0 Mar 12 00:10:57 crc kubenswrapper[4948]: I0312 00:10:57.871649 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"02288ecc-afa1-48aa-9a2d-cc78303ede88","Type":"ContainerDied","Data":"805fd24d2606a526eaa0631c6bb97d4b0bc682d491f077df0e5adc45ab3b8b25"} Mar 12 00:10:57 crc kubenswrapper[4948]: I0312 00:10:57.873283 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5488b469ff-xfnxl" event={"ID":"850a971c-3fcf-428e-8e1a-2a3a22e4afa2","Type":"ContainerStarted","Data":"bcdda0de679e88c0a984a355d469775af6d1dab8d72bc819137e828c742ef953"} Mar 12 00:10:57 crc kubenswrapper[4948]: I0312 00:10:57.873330 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5488b469ff-xfnxl" event={"ID":"850a971c-3fcf-428e-8e1a-2a3a22e4afa2","Type":"ContainerStarted","Data":"932f3f1fa8746f8a5ab5b1d0879e755ac7be65c516e805d635c85523a809a96c"} Mar 12 00:10:57 crc kubenswrapper[4948]: I0312 00:10:57.874516 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5488b469ff-xfnxl" Mar 12 00:10:57 crc kubenswrapper[4948]: I0312 00:10:57.879211 4948 generic.go:334] "Generic (PLEG): container finished" podID="f8cf7ee1-e662-4afd-9e08-112171677c22" containerID="53375f01d7cb1d105553e223936295bb8066f47dbacf62b87c6fd8cc659f2202" exitCode=0 Mar 12 00:10:57 crc kubenswrapper[4948]: I0312 00:10:57.879271 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zc9zh" event={"ID":"f8cf7ee1-e662-4afd-9e08-112171677c22","Type":"ContainerDied","Data":"53375f01d7cb1d105553e223936295bb8066f47dbacf62b87c6fd8cc659f2202"} Mar 12 00:10:57 crc kubenswrapper[4948]: I0312 00:10:57.880078 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5488b469ff-xfnxl" Mar 12 00:10:57 crc kubenswrapper[4948]: I0312 00:10:57.881246 4948 generic.go:334] "Generic (PLEG): container finished" podID="8a6ce79a-afdb-4e21-9e24-8de6924279bf" containerID="f671d445fb2b1163a7bd29f7448a3d3c3956cda4379bfe37ce56af3269fe0c79" exitCode=0 Mar 12 00:10:57 crc kubenswrapper[4948]: I0312 00:10:57.881315 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-btd4j" event={"ID":"8a6ce79a-afdb-4e21-9e24-8de6924279bf","Type":"ContainerDied","Data":"f671d445fb2b1163a7bd29f7448a3d3c3956cda4379bfe37ce56af3269fe0c79"} Mar 12 00:10:57 crc kubenswrapper[4948]: I0312 00:10:57.884293 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6w6fq" event={"ID":"529630b8-cc23-44b7-bafd-6ae1313877d1","Type":"ContainerStarted","Data":"5ed67fc29cf834f1e10fc0bf2dfcb81ae013ba954f4ebeaf9e1f55017ff9e715"} Mar 12 00:10:57 crc kubenswrapper[4948]: I0312 00:10:57.887038 4948 generic.go:334] "Generic (PLEG): container finished" podID="44d3c29e-5c1e-41be-ae1a-8d1e5f9e39c7" containerID="0a0c38f4d0f9ca9157026aed00d883062d2184f3d7694889696a5eae1fcc82b2" exitCode=0 Mar 12 00:10:57 crc kubenswrapper[4948]: I0312 00:10:57.887104 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"44d3c29e-5c1e-41be-ae1a-8d1e5f9e39c7","Type":"ContainerDied","Data":"0a0c38f4d0f9ca9157026aed00d883062d2184f3d7694889696a5eae1fcc82b2"} Mar 12 00:10:57 crc kubenswrapper[4948]: I0312 00:10:57.894524 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-74f85d4774-vqx4h" event={"ID":"6ddbb31b-e927-4588-94f2-c2ffc5b4e604","Type":"ContainerStarted","Data":"f4c9914759f108e7d0376c92af4b58e3b4d8eed6732a19a7b1cb450765437001"} Mar 12 00:10:57 crc kubenswrapper[4948]: I0312 00:10:57.894571 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-74f85d4774-vqx4h" event={"ID":"6ddbb31b-e927-4588-94f2-c2ffc5b4e604","Type":"ContainerStarted","Data":"863403b3b5a934359a7de9563e1ba0d16821bf9ffc8a5e3a2907303c7758992c"} Mar 12 00:10:57 crc kubenswrapper[4948]: I0312 00:10:57.895046 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-74f85d4774-vqx4h" Mar 12 00:10:57 crc kubenswrapper[4948]: I0312 00:10:57.902263 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-74f85d4774-vqx4h" Mar 12 00:10:58 crc kubenswrapper[4948]: I0312 00:10:58.020006 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-74f85d4774-vqx4h" podStartSLOduration=17.019976417 podStartE2EDuration="17.019976417s" podCreationTimestamp="2026-03-12 00:10:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:10:58.018915312 +0000 UTC m=+257.474519050" watchObservedRunningTime="2026-03-12 00:10:58.019976417 +0000 UTC m=+257.475580155" Mar 12 00:10:58 crc kubenswrapper[4948]: I0312 00:10:58.022176 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5488b469ff-xfnxl" podStartSLOduration=17.022164509 podStartE2EDuration="17.022164509s" podCreationTimestamp="2026-03-12 00:10:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:10:58.001032535 +0000 UTC m=+257.456636283" watchObservedRunningTime="2026-03-12 00:10:58.022164509 +0000 UTC m=+257.477768247" Mar 12 00:10:58 crc kubenswrapper[4948]: I0312 00:10:58.473737 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operator-lifecycle-manager/package-server-manager-789f6589d5-jlrln" Mar 12 00:10:58 crc kubenswrapper[4948]: I0312 00:10:58.903076 4948 generic.go:334] "Generic (PLEG): container finished" podID="529630b8-cc23-44b7-bafd-6ae1313877d1" containerID="5ed67fc29cf834f1e10fc0bf2dfcb81ae013ba954f4ebeaf9e1f55017ff9e715" exitCode=0 Mar 12 00:10:58 crc kubenswrapper[4948]: I0312 00:10:58.905113 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6w6fq" event={"ID":"529630b8-cc23-44b7-bafd-6ae1313877d1","Type":"ContainerDied","Data":"5ed67fc29cf834f1e10fc0bf2dfcb81ae013ba954f4ebeaf9e1f55017ff9e715"} Mar 12 00:10:59 crc kubenswrapper[4948]: I0312 00:10:59.194323 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 12 00:10:59 crc kubenswrapper[4948]: I0312 00:10:59.199025 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 12 00:10:59 crc kubenswrapper[4948]: I0312 00:10:59.315664 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/44d3c29e-5c1e-41be-ae1a-8d1e5f9e39c7-kubelet-dir\") pod \"44d3c29e-5c1e-41be-ae1a-8d1e5f9e39c7\" (UID: \"44d3c29e-5c1e-41be-ae1a-8d1e5f9e39c7\") " Mar 12 00:10:59 crc kubenswrapper[4948]: I0312 00:10:59.315774 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/02288ecc-afa1-48aa-9a2d-cc78303ede88-kubelet-dir\") pod \"02288ecc-afa1-48aa-9a2d-cc78303ede88\" (UID: \"02288ecc-afa1-48aa-9a2d-cc78303ede88\") " Mar 12 00:10:59 crc kubenswrapper[4948]: I0312 00:10:59.315846 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/02288ecc-afa1-48aa-9a2d-cc78303ede88-kube-api-access\") pod \"02288ecc-afa1-48aa-9a2d-cc78303ede88\" (UID: \"02288ecc-afa1-48aa-9a2d-cc78303ede88\") " Mar 12 00:10:59 crc kubenswrapper[4948]: I0312 00:10:59.315868 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/44d3c29e-5c1e-41be-ae1a-8d1e5f9e39c7-kube-api-access\") pod \"44d3c29e-5c1e-41be-ae1a-8d1e5f9e39c7\" (UID: \"44d3c29e-5c1e-41be-ae1a-8d1e5f9e39c7\") " Mar 12 00:10:59 crc kubenswrapper[4948]: I0312 00:10:59.316352 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/44d3c29e-5c1e-41be-ae1a-8d1e5f9e39c7-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "44d3c29e-5c1e-41be-ae1a-8d1e5f9e39c7" (UID: "44d3c29e-5c1e-41be-ae1a-8d1e5f9e39c7"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 00:10:59 crc kubenswrapper[4948]: I0312 00:10:59.316415 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/02288ecc-afa1-48aa-9a2d-cc78303ede88-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "02288ecc-afa1-48aa-9a2d-cc78303ede88" (UID: "02288ecc-afa1-48aa-9a2d-cc78303ede88"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 00:10:59 crc kubenswrapper[4948]: I0312 00:10:59.321542 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44d3c29e-5c1e-41be-ae1a-8d1e5f9e39c7-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "44d3c29e-5c1e-41be-ae1a-8d1e5f9e39c7" (UID: "44d3c29e-5c1e-41be-ae1a-8d1e5f9e39c7"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:10:59 crc kubenswrapper[4948]: I0312 00:10:59.335750 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/02288ecc-afa1-48aa-9a2d-cc78303ede88-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "02288ecc-afa1-48aa-9a2d-cc78303ede88" (UID: "02288ecc-afa1-48aa-9a2d-cc78303ede88"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:10:59 crc kubenswrapper[4948]: I0312 00:10:59.417764 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/02288ecc-afa1-48aa-9a2d-cc78303ede88-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 12 00:10:59 crc kubenswrapper[4948]: I0312 00:10:59.417816 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/44d3c29e-5c1e-41be-ae1a-8d1e5f9e39c7-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 12 00:10:59 crc kubenswrapper[4948]: I0312 00:10:59.417837 4948 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/44d3c29e-5c1e-41be-ae1a-8d1e5f9e39c7-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 12 00:10:59 crc kubenswrapper[4948]: I0312 00:10:59.417857 4948 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/02288ecc-afa1-48aa-9a2d-cc78303ede88-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 12 00:10:59 crc kubenswrapper[4948]: I0312 00:10:59.886951 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 12 00:10:59 crc kubenswrapper[4948]: E0312 00:10:59.887583 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="02288ecc-afa1-48aa-9a2d-cc78303ede88" containerName="pruner" Mar 12 00:10:59 crc kubenswrapper[4948]: I0312 00:10:59.887595 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="02288ecc-afa1-48aa-9a2d-cc78303ede88" containerName="pruner" Mar 12 00:10:59 crc kubenswrapper[4948]: E0312 00:10:59.887609 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44d3c29e-5c1e-41be-ae1a-8d1e5f9e39c7" containerName="pruner" Mar 12 00:10:59 crc kubenswrapper[4948]: I0312 00:10:59.887616 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="44d3c29e-5c1e-41be-ae1a-8d1e5f9e39c7" containerName="pruner" Mar 12 00:10:59 crc kubenswrapper[4948]: I0312 00:10:59.887740 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="02288ecc-afa1-48aa-9a2d-cc78303ede88" containerName="pruner" Mar 12 00:10:59 crc kubenswrapper[4948]: I0312 00:10:59.887751 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="44d3c29e-5c1e-41be-ae1a-8d1e5f9e39c7" containerName="pruner" Mar 12 00:10:59 crc kubenswrapper[4948]: I0312 00:10:59.888189 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 12 00:10:59 crc kubenswrapper[4948]: I0312 00:10:59.905583 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 12 00:10:59 crc kubenswrapper[4948]: I0312 00:10:59.926818 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7aa3a362-6bdc-409f-8341-6c0c494f163c-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"7aa3a362-6bdc-409f-8341-6c0c494f163c\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 12 00:10:59 crc kubenswrapper[4948]: I0312 00:10:59.926910 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7aa3a362-6bdc-409f-8341-6c0c494f163c-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"7aa3a362-6bdc-409f-8341-6c0c494f163c\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 12 00:10:59 crc kubenswrapper[4948]: I0312 00:10:59.931564 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-8-crc" Mar 12 00:10:59 crc kubenswrapper[4948]: I0312 00:10:59.932106 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-8-crc" event={"ID":"02288ecc-afa1-48aa-9a2d-cc78303ede88","Type":"ContainerDied","Data":"2046154f8b6ca6a025f01d7e1a0015386ef3503c4d31b4ee7e6e681a637be797"} Mar 12 00:10:59 crc kubenswrapper[4948]: I0312 00:10:59.932151 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="2046154f8b6ca6a025f01d7e1a0015386ef3503c4d31b4ee7e6e681a637be797" Mar 12 00:10:59 crc kubenswrapper[4948]: I0312 00:10:59.940884 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/revision-pruner-9-crc" event={"ID":"44d3c29e-5c1e-41be-ae1a-8d1e5f9e39c7","Type":"ContainerDied","Data":"5e78897f10b9b3c3672515acd885ca4d0edda2efc255ea97ead64bc5ba2f6aca"} Mar 12 00:10:59 crc kubenswrapper[4948]: I0312 00:10:59.940945 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5e78897f10b9b3c3672515acd885ca4d0edda2efc255ea97ead64bc5ba2f6aca" Mar 12 00:10:59 crc kubenswrapper[4948]: I0312 00:10:59.940968 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-controller-manager/revision-pruner-9-crc" Mar 12 00:11:00 crc kubenswrapper[4948]: I0312 00:11:00.028592 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7aa3a362-6bdc-409f-8341-6c0c494f163c-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"7aa3a362-6bdc-409f-8341-6c0c494f163c\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 12 00:11:00 crc kubenswrapper[4948]: I0312 00:11:00.028700 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7aa3a362-6bdc-409f-8341-6c0c494f163c-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"7aa3a362-6bdc-409f-8341-6c0c494f163c\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 12 00:11:00 crc kubenswrapper[4948]: I0312 00:11:00.029134 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7aa3a362-6bdc-409f-8341-6c0c494f163c-kubelet-dir\") pod \"revision-pruner-9-crc\" (UID: \"7aa3a362-6bdc-409f-8341-6c0c494f163c\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 12 00:11:00 crc kubenswrapper[4948]: I0312 00:11:00.045081 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7aa3a362-6bdc-409f-8341-6c0c494f163c-kube-api-access\") pod \"revision-pruner-9-crc\" (UID: \"7aa3a362-6bdc-409f-8341-6c0c494f163c\") " pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 12 00:11:00 crc kubenswrapper[4948]: I0312 00:11:00.293984 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 12 00:11:00 crc kubenswrapper[4948]: I0312 00:11:00.689280 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/revision-pruner-9-crc"] Mar 12 00:11:00 crc kubenswrapper[4948]: I0312 00:11:00.980850 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zc9zh" event={"ID":"f8cf7ee1-e662-4afd-9e08-112171677c22","Type":"ContainerStarted","Data":"326ebf4817bc847d36148437a6dcda1508f716edd2d55a3be608cafdfbfb5f7e"} Mar 12 00:11:00 crc kubenswrapper[4948]: I0312 00:11:00.983207 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-btd4j" event={"ID":"8a6ce79a-afdb-4e21-9e24-8de6924279bf","Type":"ContainerStarted","Data":"b388539342b4cfeef5eebb31026e89abad176dae01f1b209244ea76136260c19"} Mar 12 00:11:00 crc kubenswrapper[4948]: I0312 00:11:00.985800 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6w6fq" event={"ID":"529630b8-cc23-44b7-bafd-6ae1313877d1","Type":"ContainerStarted","Data":"a40fecd469f3d505800497aa8b7497e5f3c26dbab5363fa4708bc6ad58b76789"} Mar 12 00:11:00 crc kubenswrapper[4948]: I0312 00:11:00.989240 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bsmcp" event={"ID":"44212e8e-c675-4643-a752-997be5f8e980","Type":"ContainerStarted","Data":"d012b13a227c5089a7a52d6cf33547053badf64ce2fbf7f65be8dec74ef2954f"} Mar 12 00:11:00 crc kubenswrapper[4948]: I0312 00:11:00.991016 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"7aa3a362-6bdc-409f-8341-6c0c494f163c","Type":"ContainerStarted","Data":"0e63b2912cc5d71a8361c36427c660ade2286a4d4193a0aef26a13c2aaa5d183"} Mar 12 00:11:01 crc kubenswrapper[4948]: I0312 00:11:01.001230 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-zc9zh" podStartSLOduration=4.140827949 podStartE2EDuration="35.001211511s" podCreationTimestamp="2026-03-12 00:10:26 +0000 UTC" firstStartedPulling="2026-03-12 00:10:28.697729063 +0000 UTC m=+228.153332801" lastFinishedPulling="2026-03-12 00:10:59.558112595 +0000 UTC m=+259.013716363" observedRunningTime="2026-03-12 00:11:00.998276072 +0000 UTC m=+260.453879810" watchObservedRunningTime="2026-03-12 00:11:01.001211511 +0000 UTC m=+260.456815259" Mar 12 00:11:01 crc kubenswrapper[4948]: I0312 00:11:01.019178 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-bsmcp" podStartSLOduration=2.908894347 podStartE2EDuration="34.019157291s" podCreationTimestamp="2026-03-12 00:10:27 +0000 UTC" firstStartedPulling="2026-03-12 00:10:28.697728083 +0000 UTC m=+228.153331821" lastFinishedPulling="2026-03-12 00:10:59.807990997 +0000 UTC m=+259.263594765" observedRunningTime="2026-03-12 00:11:01.01569792 +0000 UTC m=+260.471301668" watchObservedRunningTime="2026-03-12 00:11:01.019157291 +0000 UTC m=+260.474761049" Mar 12 00:11:01 crc kubenswrapper[4948]: I0312 00:11:01.039273 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-btd4j" podStartSLOduration=12.399307812 podStartE2EDuration="34.039255211s" podCreationTimestamp="2026-03-12 00:10:27 +0000 UTC" firstStartedPulling="2026-03-12 00:10:38.125064192 +0000 UTC m=+237.580667930" lastFinishedPulling="2026-03-12 00:10:59.765011561 +0000 UTC m=+259.220615329" observedRunningTime="2026-03-12 00:11:01.038335759 +0000 UTC m=+260.493939507" watchObservedRunningTime="2026-03-12 00:11:01.039255211 +0000 UTC m=+260.494858959" Mar 12 00:11:01 crc kubenswrapper[4948]: I0312 00:11:01.056107 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-6w6fq" podStartSLOduration=30.038239534 podStartE2EDuration="33.056094265s" podCreationTimestamp="2026-03-12 00:10:28 +0000 UTC" firstStartedPulling="2026-03-12 00:10:56.848148892 +0000 UTC m=+256.303752630" lastFinishedPulling="2026-03-12 00:10:59.866003623 +0000 UTC m=+259.321607361" observedRunningTime="2026-03-12 00:11:01.055067791 +0000 UTC m=+260.510671539" watchObservedRunningTime="2026-03-12 00:11:01.056094265 +0000 UTC m=+260.511698003" Mar 12 00:11:01 crc kubenswrapper[4948]: I0312 00:11:01.302546 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-74f85d4774-vqx4h"] Mar 12 00:11:01 crc kubenswrapper[4948]: I0312 00:11:01.302760 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-74f85d4774-vqx4h" podUID="6ddbb31b-e927-4588-94f2-c2ffc5b4e604" containerName="controller-manager" containerID="cri-o://f4c9914759f108e7d0376c92af4b58e3b4d8eed6732a19a7b1cb450765437001" gracePeriod=30 Mar 12 00:11:01 crc kubenswrapper[4948]: I0312 00:11:01.424246 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5488b469ff-xfnxl"] Mar 12 00:11:01 crc kubenswrapper[4948]: I0312 00:11:01.424478 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-5488b469ff-xfnxl" podUID="850a971c-3fcf-428e-8e1a-2a3a22e4afa2" containerName="route-controller-manager" containerID="cri-o://bcdda0de679e88c0a984a355d469775af6d1dab8d72bc819137e828c742ef953" gracePeriod=30 Mar 12 00:11:01 crc kubenswrapper[4948]: I0312 00:11:01.815832 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-74f85d4774-vqx4h" Mar 12 00:11:01 crc kubenswrapper[4948]: I0312 00:11:01.819225 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5488b469ff-xfnxl" Mar 12 00:11:01 crc kubenswrapper[4948]: I0312 00:11:01.982875 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6ddbb31b-e927-4588-94f2-c2ffc5b4e604-proxy-ca-bundles\") pod \"6ddbb31b-e927-4588-94f2-c2ffc5b4e604\" (UID: \"6ddbb31b-e927-4588-94f2-c2ffc5b4e604\") " Mar 12 00:11:01 crc kubenswrapper[4948]: I0312 00:11:01.982961 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ddbb31b-e927-4588-94f2-c2ffc5b4e604-config\") pod \"6ddbb31b-e927-4588-94f2-c2ffc5b4e604\" (UID: \"6ddbb31b-e927-4588-94f2-c2ffc5b4e604\") " Mar 12 00:11:01 crc kubenswrapper[4948]: I0312 00:11:01.982982 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/850a971c-3fcf-428e-8e1a-2a3a22e4afa2-config\") pod \"850a971c-3fcf-428e-8e1a-2a3a22e4afa2\" (UID: \"850a971c-3fcf-428e-8e1a-2a3a22e4afa2\") " Mar 12 00:11:01 crc kubenswrapper[4948]: I0312 00:11:01.983011 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8j7s6\" (UniqueName: \"kubernetes.io/projected/850a971c-3fcf-428e-8e1a-2a3a22e4afa2-kube-api-access-8j7s6\") pod \"850a971c-3fcf-428e-8e1a-2a3a22e4afa2\" (UID: \"850a971c-3fcf-428e-8e1a-2a3a22e4afa2\") " Mar 12 00:11:01 crc kubenswrapper[4948]: I0312 00:11:01.983037 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-64qjf\" (UniqueName: \"kubernetes.io/projected/6ddbb31b-e927-4588-94f2-c2ffc5b4e604-kube-api-access-64qjf\") pod \"6ddbb31b-e927-4588-94f2-c2ffc5b4e604\" (UID: \"6ddbb31b-e927-4588-94f2-c2ffc5b4e604\") " Mar 12 00:11:01 crc kubenswrapper[4948]: I0312 00:11:01.983060 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/850a971c-3fcf-428e-8e1a-2a3a22e4afa2-serving-cert\") pod \"850a971c-3fcf-428e-8e1a-2a3a22e4afa2\" (UID: \"850a971c-3fcf-428e-8e1a-2a3a22e4afa2\") " Mar 12 00:11:01 crc kubenswrapper[4948]: I0312 00:11:01.983085 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6ddbb31b-e927-4588-94f2-c2ffc5b4e604-client-ca\") pod \"6ddbb31b-e927-4588-94f2-c2ffc5b4e604\" (UID: \"6ddbb31b-e927-4588-94f2-c2ffc5b4e604\") " Mar 12 00:11:01 crc kubenswrapper[4948]: I0312 00:11:01.983100 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6ddbb31b-e927-4588-94f2-c2ffc5b4e604-serving-cert\") pod \"6ddbb31b-e927-4588-94f2-c2ffc5b4e604\" (UID: \"6ddbb31b-e927-4588-94f2-c2ffc5b4e604\") " Mar 12 00:11:01 crc kubenswrapper[4948]: I0312 00:11:01.983139 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/850a971c-3fcf-428e-8e1a-2a3a22e4afa2-client-ca\") pod \"850a971c-3fcf-428e-8e1a-2a3a22e4afa2\" (UID: \"850a971c-3fcf-428e-8e1a-2a3a22e4afa2\") " Mar 12 00:11:01 crc kubenswrapper[4948]: I0312 00:11:01.983715 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/850a971c-3fcf-428e-8e1a-2a3a22e4afa2-client-ca" (OuterVolumeSpecName: "client-ca") pod "850a971c-3fcf-428e-8e1a-2a3a22e4afa2" (UID: "850a971c-3fcf-428e-8e1a-2a3a22e4afa2"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:11:01 crc kubenswrapper[4948]: I0312 00:11:01.983741 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/850a971c-3fcf-428e-8e1a-2a3a22e4afa2-config" (OuterVolumeSpecName: "config") pod "850a971c-3fcf-428e-8e1a-2a3a22e4afa2" (UID: "850a971c-3fcf-428e-8e1a-2a3a22e4afa2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:11:01 crc kubenswrapper[4948]: I0312 00:11:01.983760 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ddbb31b-e927-4588-94f2-c2ffc5b4e604-config" (OuterVolumeSpecName: "config") pod "6ddbb31b-e927-4588-94f2-c2ffc5b4e604" (UID: "6ddbb31b-e927-4588-94f2-c2ffc5b4e604"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:11:01 crc kubenswrapper[4948]: I0312 00:11:01.983836 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ddbb31b-e927-4588-94f2-c2ffc5b4e604-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "6ddbb31b-e927-4588-94f2-c2ffc5b4e604" (UID: "6ddbb31b-e927-4588-94f2-c2ffc5b4e604"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:11:01 crc kubenswrapper[4948]: I0312 00:11:01.984002 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/6ddbb31b-e927-4588-94f2-c2ffc5b4e604-client-ca" (OuterVolumeSpecName: "client-ca") pod "6ddbb31b-e927-4588-94f2-c2ffc5b4e604" (UID: "6ddbb31b-e927-4588-94f2-c2ffc5b4e604"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:11:01 crc kubenswrapper[4948]: I0312 00:11:01.988558 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/6ddbb31b-e927-4588-94f2-c2ffc5b4e604-kube-api-access-64qjf" (OuterVolumeSpecName: "kube-api-access-64qjf") pod "6ddbb31b-e927-4588-94f2-c2ffc5b4e604" (UID: "6ddbb31b-e927-4588-94f2-c2ffc5b4e604"). InnerVolumeSpecName "kube-api-access-64qjf". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:11:01 crc kubenswrapper[4948]: I0312 00:11:01.988643 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/850a971c-3fcf-428e-8e1a-2a3a22e4afa2-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "850a971c-3fcf-428e-8e1a-2a3a22e4afa2" (UID: "850a971c-3fcf-428e-8e1a-2a3a22e4afa2"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:11:01 crc kubenswrapper[4948]: I0312 00:11:01.992205 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/850a971c-3fcf-428e-8e1a-2a3a22e4afa2-kube-api-access-8j7s6" (OuterVolumeSpecName: "kube-api-access-8j7s6") pod "850a971c-3fcf-428e-8e1a-2a3a22e4afa2" (UID: "850a971c-3fcf-428e-8e1a-2a3a22e4afa2"). InnerVolumeSpecName "kube-api-access-8j7s6". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:11:01 crc kubenswrapper[4948]: I0312 00:11:01.992242 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/6ddbb31b-e927-4588-94f2-c2ffc5b4e604-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "6ddbb31b-e927-4588-94f2-c2ffc5b4e604" (UID: "6ddbb31b-e927-4588-94f2-c2ffc5b4e604"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:11:01 crc kubenswrapper[4948]: I0312 00:11:01.996588 4948 generic.go:334] "Generic (PLEG): container finished" podID="6ddbb31b-e927-4588-94f2-c2ffc5b4e604" containerID="f4c9914759f108e7d0376c92af4b58e3b4d8eed6732a19a7b1cb450765437001" exitCode=0 Mar 12 00:11:01 crc kubenswrapper[4948]: I0312 00:11:01.996644 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-74f85d4774-vqx4h" event={"ID":"6ddbb31b-e927-4588-94f2-c2ffc5b4e604","Type":"ContainerDied","Data":"f4c9914759f108e7d0376c92af4b58e3b4d8eed6732a19a7b1cb450765437001"} Mar 12 00:11:01 crc kubenswrapper[4948]: I0312 00:11:01.996674 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-74f85d4774-vqx4h" event={"ID":"6ddbb31b-e927-4588-94f2-c2ffc5b4e604","Type":"ContainerDied","Data":"863403b3b5a934359a7de9563e1ba0d16821bf9ffc8a5e3a2907303c7758992c"} Mar 12 00:11:01 crc kubenswrapper[4948]: I0312 00:11:01.996689 4948 scope.go:117] "RemoveContainer" containerID="f4c9914759f108e7d0376c92af4b58e3b4d8eed6732a19a7b1cb450765437001" Mar 12 00:11:01 crc kubenswrapper[4948]: I0312 00:11:01.996780 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-74f85d4774-vqx4h" Mar 12 00:11:02 crc kubenswrapper[4948]: I0312 00:11:02.001346 4948 generic.go:334] "Generic (PLEG): container finished" podID="850a971c-3fcf-428e-8e1a-2a3a22e4afa2" containerID="bcdda0de679e88c0a984a355d469775af6d1dab8d72bc819137e828c742ef953" exitCode=0 Mar 12 00:11:02 crc kubenswrapper[4948]: I0312 00:11:02.001423 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5488b469ff-xfnxl" event={"ID":"850a971c-3fcf-428e-8e1a-2a3a22e4afa2","Type":"ContainerDied","Data":"bcdda0de679e88c0a984a355d469775af6d1dab8d72bc819137e828c742ef953"} Mar 12 00:11:02 crc kubenswrapper[4948]: I0312 00:11:02.001456 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5488b469ff-xfnxl" event={"ID":"850a971c-3fcf-428e-8e1a-2a3a22e4afa2","Type":"ContainerDied","Data":"932f3f1fa8746f8a5ab5b1d0879e755ac7be65c516e805d635c85523a809a96c"} Mar 12 00:11:02 crc kubenswrapper[4948]: I0312 00:11:02.001519 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5488b469ff-xfnxl" Mar 12 00:11:02 crc kubenswrapper[4948]: I0312 00:11:02.006871 4948 generic.go:334] "Generic (PLEG): container finished" podID="7aa3a362-6bdc-409f-8341-6c0c494f163c" containerID="c9c918674a1e7f07a0df009963e952109472f794d059dd135222166b1111d11c" exitCode=0 Mar 12 00:11:02 crc kubenswrapper[4948]: I0312 00:11:02.007578 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"7aa3a362-6bdc-409f-8341-6c0c494f163c","Type":"ContainerDied","Data":"c9c918674a1e7f07a0df009963e952109472f794d059dd135222166b1111d11c"} Mar 12 00:11:02 crc kubenswrapper[4948]: I0312 00:11:02.013141 4948 scope.go:117] "RemoveContainer" containerID="f4c9914759f108e7d0376c92af4b58e3b4d8eed6732a19a7b1cb450765437001" Mar 12 00:11:02 crc kubenswrapper[4948]: E0312 00:11:02.013482 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"f4c9914759f108e7d0376c92af4b58e3b4d8eed6732a19a7b1cb450765437001\": container with ID starting with f4c9914759f108e7d0376c92af4b58e3b4d8eed6732a19a7b1cb450765437001 not found: ID does not exist" containerID="f4c9914759f108e7d0376c92af4b58e3b4d8eed6732a19a7b1cb450765437001" Mar 12 00:11:02 crc kubenswrapper[4948]: I0312 00:11:02.013524 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"f4c9914759f108e7d0376c92af4b58e3b4d8eed6732a19a7b1cb450765437001"} err="failed to get container status \"f4c9914759f108e7d0376c92af4b58e3b4d8eed6732a19a7b1cb450765437001\": rpc error: code = NotFound desc = could not find container \"f4c9914759f108e7d0376c92af4b58e3b4d8eed6732a19a7b1cb450765437001\": container with ID starting with f4c9914759f108e7d0376c92af4b58e3b4d8eed6732a19a7b1cb450765437001 not found: ID does not exist" Mar 12 00:11:02 crc kubenswrapper[4948]: I0312 00:11:02.013554 4948 scope.go:117] "RemoveContainer" containerID="bcdda0de679e88c0a984a355d469775af6d1dab8d72bc819137e828c742ef953" Mar 12 00:11:02 crc kubenswrapper[4948]: I0312 00:11:02.031975 4948 scope.go:117] "RemoveContainer" containerID="bcdda0de679e88c0a984a355d469775af6d1dab8d72bc819137e828c742ef953" Mar 12 00:11:02 crc kubenswrapper[4948]: E0312 00:11:02.032657 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bcdda0de679e88c0a984a355d469775af6d1dab8d72bc819137e828c742ef953\": container with ID starting with bcdda0de679e88c0a984a355d469775af6d1dab8d72bc819137e828c742ef953 not found: ID does not exist" containerID="bcdda0de679e88c0a984a355d469775af6d1dab8d72bc819137e828c742ef953" Mar 12 00:11:02 crc kubenswrapper[4948]: I0312 00:11:02.032715 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bcdda0de679e88c0a984a355d469775af6d1dab8d72bc819137e828c742ef953"} err="failed to get container status \"bcdda0de679e88c0a984a355d469775af6d1dab8d72bc819137e828c742ef953\": rpc error: code = NotFound desc = could not find container \"bcdda0de679e88c0a984a355d469775af6d1dab8d72bc819137e828c742ef953\": container with ID starting with bcdda0de679e88c0a984a355d469775af6d1dab8d72bc819137e828c742ef953 not found: ID does not exist" Mar 12 00:11:02 crc kubenswrapper[4948]: I0312 00:11:02.037206 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-74f85d4774-vqx4h"] Mar 12 00:11:02 crc kubenswrapper[4948]: I0312 00:11:02.039116 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-74f85d4774-vqx4h"] Mar 12 00:11:02 crc kubenswrapper[4948]: I0312 00:11:02.048910 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5488b469ff-xfnxl"] Mar 12 00:11:02 crc kubenswrapper[4948]: I0312 00:11:02.055558 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5488b469ff-xfnxl"] Mar 12 00:11:02 crc kubenswrapper[4948]: I0312 00:11:02.084372 4948 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/850a971c-3fcf-428e-8e1a-2a3a22e4afa2-client-ca\") on node \"crc\" DevicePath \"\"" Mar 12 00:11:02 crc kubenswrapper[4948]: I0312 00:11:02.084409 4948 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/6ddbb31b-e927-4588-94f2-c2ffc5b4e604-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 12 00:11:02 crc kubenswrapper[4948]: I0312 00:11:02.084421 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/6ddbb31b-e927-4588-94f2-c2ffc5b4e604-config\") on node \"crc\" DevicePath \"\"" Mar 12 00:11:02 crc kubenswrapper[4948]: I0312 00:11:02.084429 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/850a971c-3fcf-428e-8e1a-2a3a22e4afa2-config\") on node \"crc\" DevicePath \"\"" Mar 12 00:11:02 crc kubenswrapper[4948]: I0312 00:11:02.084438 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8j7s6\" (UniqueName: \"kubernetes.io/projected/850a971c-3fcf-428e-8e1a-2a3a22e4afa2-kube-api-access-8j7s6\") on node \"crc\" DevicePath \"\"" Mar 12 00:11:02 crc kubenswrapper[4948]: I0312 00:11:02.084447 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-64qjf\" (UniqueName: \"kubernetes.io/projected/6ddbb31b-e927-4588-94f2-c2ffc5b4e604-kube-api-access-64qjf\") on node \"crc\" DevicePath \"\"" Mar 12 00:11:02 crc kubenswrapper[4948]: I0312 00:11:02.084461 4948 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/850a971c-3fcf-428e-8e1a-2a3a22e4afa2-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 00:11:02 crc kubenswrapper[4948]: I0312 00:11:02.084468 4948 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/6ddbb31b-e927-4588-94f2-c2ffc5b4e604-client-ca\") on node \"crc\" DevicePath \"\"" Mar 12 00:11:02 crc kubenswrapper[4948]: I0312 00:11:02.084476 4948 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/6ddbb31b-e927-4588-94f2-c2ffc5b4e604-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 00:11:02 crc kubenswrapper[4948]: I0312 00:11:02.464245 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-network-diagnostics/network-check-target-xd92c" Mar 12 00:11:02 crc kubenswrapper[4948]: I0312 00:11:02.659825 4948 patch_prober.go:28] interesting pod/controller-manager-74f85d4774-vqx4h container/controller-manager namespace/openshift-controller-manager: Readiness probe status=failure output="Get \"https://10.217.0.58:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" start-of-body= Mar 12 00:11:02 crc kubenswrapper[4948]: I0312 00:11:02.660247 4948 prober.go:107] "Probe failed" probeType="Readiness" pod="openshift-controller-manager/controller-manager-74f85d4774-vqx4h" podUID="6ddbb31b-e927-4588-94f2-c2ffc5b4e604" containerName="controller-manager" probeResult="failure" output="Get \"https://10.217.0.58:8443/healthz\": net/http: request canceled while waiting for connection (Client.Timeout exceeded while awaiting headers)" Mar 12 00:11:03 crc kubenswrapper[4948]: I0312 00:11:03.251323 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-7f5b4654f4-jnbsn"] Mar 12 00:11:03 crc kubenswrapper[4948]: E0312 00:11:03.251633 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="850a971c-3fcf-428e-8e1a-2a3a22e4afa2" containerName="route-controller-manager" Mar 12 00:11:03 crc kubenswrapper[4948]: I0312 00:11:03.251649 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="850a971c-3fcf-428e-8e1a-2a3a22e4afa2" containerName="route-controller-manager" Mar 12 00:11:03 crc kubenswrapper[4948]: E0312 00:11:03.251660 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="6ddbb31b-e927-4588-94f2-c2ffc5b4e604" containerName="controller-manager" Mar 12 00:11:03 crc kubenswrapper[4948]: I0312 00:11:03.251669 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="6ddbb31b-e927-4588-94f2-c2ffc5b4e604" containerName="controller-manager" Mar 12 00:11:03 crc kubenswrapper[4948]: I0312 00:11:03.251792 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="850a971c-3fcf-428e-8e1a-2a3a22e4afa2" containerName="route-controller-manager" Mar 12 00:11:03 crc kubenswrapper[4948]: I0312 00:11:03.251809 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="6ddbb31b-e927-4588-94f2-c2ffc5b4e604" containerName="controller-manager" Mar 12 00:11:03 crc kubenswrapper[4948]: I0312 00:11:03.252272 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7f5b4654f4-jnbsn" Mar 12 00:11:03 crc kubenswrapper[4948]: I0312 00:11:03.261955 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-66dcfbd4c4-7wvzp"] Mar 12 00:11:03 crc kubenswrapper[4948]: I0312 00:11:03.264633 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-66dcfbd4c4-7wvzp" Mar 12 00:11:03 crc kubenswrapper[4948]: I0312 00:11:03.272118 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 12 00:11:03 crc kubenswrapper[4948]: I0312 00:11:03.274785 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 12 00:11:03 crc kubenswrapper[4948]: I0312 00:11:03.274935 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 12 00:11:03 crc kubenswrapper[4948]: I0312 00:11:03.275733 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 12 00:11:03 crc kubenswrapper[4948]: I0312 00:11:03.275841 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 12 00:11:03 crc kubenswrapper[4948]: I0312 00:11:03.275929 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 12 00:11:03 crc kubenswrapper[4948]: I0312 00:11:03.277111 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 12 00:11:03 crc kubenswrapper[4948]: I0312 00:11:03.283962 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 12 00:11:03 crc kubenswrapper[4948]: I0312 00:11:03.284590 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 12 00:11:03 crc kubenswrapper[4948]: I0312 00:11:03.284632 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 12 00:11:03 crc kubenswrapper[4948]: I0312 00:11:03.284693 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 12 00:11:03 crc kubenswrapper[4948]: I0312 00:11:03.284911 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 12 00:11:03 crc kubenswrapper[4948]: I0312 00:11:03.284917 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 12 00:11:03 crc kubenswrapper[4948]: I0312 00:11:03.284985 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 12 00:11:03 crc kubenswrapper[4948]: I0312 00:11:03.286228 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-66dcfbd4c4-7wvzp"] Mar 12 00:11:03 crc kubenswrapper[4948]: I0312 00:11:03.299687 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7f5b4654f4-jnbsn"] Mar 12 00:11:03 crc kubenswrapper[4948]: I0312 00:11:03.324784 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="6ddbb31b-e927-4588-94f2-c2ffc5b4e604" path="/var/lib/kubelet/pods/6ddbb31b-e927-4588-94f2-c2ffc5b4e604/volumes" Mar 12 00:11:03 crc kubenswrapper[4948]: I0312 00:11:03.325496 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="850a971c-3fcf-428e-8e1a-2a3a22e4afa2" path="/var/lib/kubelet/pods/850a971c-3fcf-428e-8e1a-2a3a22e4afa2/volumes" Mar 12 00:11:03 crc kubenswrapper[4948]: I0312 00:11:03.433784 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7aa3a362-6bdc-409f-8341-6c0c494f163c-kube-api-access\") pod \"7aa3a362-6bdc-409f-8341-6c0c494f163c\" (UID: \"7aa3a362-6bdc-409f-8341-6c0c494f163c\") " Mar 12 00:11:03 crc kubenswrapper[4948]: I0312 00:11:03.433828 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7aa3a362-6bdc-409f-8341-6c0c494f163c-kubelet-dir\") pod \"7aa3a362-6bdc-409f-8341-6c0c494f163c\" (UID: \"7aa3a362-6bdc-409f-8341-6c0c494f163c\") " Mar 12 00:11:03 crc kubenswrapper[4948]: I0312 00:11:03.433979 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/25bd67ae-99ad-4f13-a87f-8416e060d3b2-serving-cert\") pod \"controller-manager-7f5b4654f4-jnbsn\" (UID: \"25bd67ae-99ad-4f13-a87f-8416e060d3b2\") " pod="openshift-controller-manager/controller-manager-7f5b4654f4-jnbsn" Mar 12 00:11:03 crc kubenswrapper[4948]: I0312 00:11:03.434017 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/25bd67ae-99ad-4f13-a87f-8416e060d3b2-proxy-ca-bundles\") pod \"controller-manager-7f5b4654f4-jnbsn\" (UID: \"25bd67ae-99ad-4f13-a87f-8416e060d3b2\") " pod="openshift-controller-manager/controller-manager-7f5b4654f4-jnbsn" Mar 12 00:11:03 crc kubenswrapper[4948]: I0312 00:11:03.434056 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-b97hz\" (UniqueName: \"kubernetes.io/projected/25bd67ae-99ad-4f13-a87f-8416e060d3b2-kube-api-access-b97hz\") pod \"controller-manager-7f5b4654f4-jnbsn\" (UID: \"25bd67ae-99ad-4f13-a87f-8416e060d3b2\") " pod="openshift-controller-manager/controller-manager-7f5b4654f4-jnbsn" Mar 12 00:11:03 crc kubenswrapper[4948]: I0312 00:11:03.434077 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/23f23bd1-c3ee-4ec7-9abb-ca6d2b3d0deb-serving-cert\") pod \"route-controller-manager-66dcfbd4c4-7wvzp\" (UID: \"23f23bd1-c3ee-4ec7-9abb-ca6d2b3d0deb\") " pod="openshift-route-controller-manager/route-controller-manager-66dcfbd4c4-7wvzp" Mar 12 00:11:03 crc kubenswrapper[4948]: I0312 00:11:03.434073 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/7aa3a362-6bdc-409f-8341-6c0c494f163c-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "7aa3a362-6bdc-409f-8341-6c0c494f163c" (UID: "7aa3a362-6bdc-409f-8341-6c0c494f163c"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 00:11:03 crc kubenswrapper[4948]: I0312 00:11:03.434166 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/25bd67ae-99ad-4f13-a87f-8416e060d3b2-client-ca\") pod \"controller-manager-7f5b4654f4-jnbsn\" (UID: \"25bd67ae-99ad-4f13-a87f-8416e060d3b2\") " pod="openshift-controller-manager/controller-manager-7f5b4654f4-jnbsn" Mar 12 00:11:03 crc kubenswrapper[4948]: I0312 00:11:03.434219 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23f23bd1-c3ee-4ec7-9abb-ca6d2b3d0deb-config\") pod \"route-controller-manager-66dcfbd4c4-7wvzp\" (UID: \"23f23bd1-c3ee-4ec7-9abb-ca6d2b3d0deb\") " pod="openshift-route-controller-manager/route-controller-manager-66dcfbd4c4-7wvzp" Mar 12 00:11:03 crc kubenswrapper[4948]: I0312 00:11:03.434251 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-d9r2j\" (UniqueName: \"kubernetes.io/projected/23f23bd1-c3ee-4ec7-9abb-ca6d2b3d0deb-kube-api-access-d9r2j\") pod \"route-controller-manager-66dcfbd4c4-7wvzp\" (UID: \"23f23bd1-c3ee-4ec7-9abb-ca6d2b3d0deb\") " pod="openshift-route-controller-manager/route-controller-manager-66dcfbd4c4-7wvzp" Mar 12 00:11:03 crc kubenswrapper[4948]: I0312 00:11:03.434337 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25bd67ae-99ad-4f13-a87f-8416e060d3b2-config\") pod \"controller-manager-7f5b4654f4-jnbsn\" (UID: \"25bd67ae-99ad-4f13-a87f-8416e060d3b2\") " pod="openshift-controller-manager/controller-manager-7f5b4654f4-jnbsn" Mar 12 00:11:03 crc kubenswrapper[4948]: I0312 00:11:03.434377 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/23f23bd1-c3ee-4ec7-9abb-ca6d2b3d0deb-client-ca\") pod \"route-controller-manager-66dcfbd4c4-7wvzp\" (UID: \"23f23bd1-c3ee-4ec7-9abb-ca6d2b3d0deb\") " pod="openshift-route-controller-manager/route-controller-manager-66dcfbd4c4-7wvzp" Mar 12 00:11:03 crc kubenswrapper[4948]: I0312 00:11:03.434434 4948 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/7aa3a362-6bdc-409f-8341-6c0c494f163c-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 12 00:11:03 crc kubenswrapper[4948]: I0312 00:11:03.439215 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7aa3a362-6bdc-409f-8341-6c0c494f163c-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "7aa3a362-6bdc-409f-8341-6c0c494f163c" (UID: "7aa3a362-6bdc-409f-8341-6c0c494f163c"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:11:03 crc kubenswrapper[4948]: I0312 00:11:03.534816 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-b97hz\" (UniqueName: \"kubernetes.io/projected/25bd67ae-99ad-4f13-a87f-8416e060d3b2-kube-api-access-b97hz\") pod \"controller-manager-7f5b4654f4-jnbsn\" (UID: \"25bd67ae-99ad-4f13-a87f-8416e060d3b2\") " pod="openshift-controller-manager/controller-manager-7f5b4654f4-jnbsn" Mar 12 00:11:03 crc kubenswrapper[4948]: I0312 00:11:03.534862 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/23f23bd1-c3ee-4ec7-9abb-ca6d2b3d0deb-serving-cert\") pod \"route-controller-manager-66dcfbd4c4-7wvzp\" (UID: \"23f23bd1-c3ee-4ec7-9abb-ca6d2b3d0deb\") " pod="openshift-route-controller-manager/route-controller-manager-66dcfbd4c4-7wvzp" Mar 12 00:11:03 crc kubenswrapper[4948]: I0312 00:11:03.534898 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/25bd67ae-99ad-4f13-a87f-8416e060d3b2-client-ca\") pod \"controller-manager-7f5b4654f4-jnbsn\" (UID: \"25bd67ae-99ad-4f13-a87f-8416e060d3b2\") " pod="openshift-controller-manager/controller-manager-7f5b4654f4-jnbsn" Mar 12 00:11:03 crc kubenswrapper[4948]: I0312 00:11:03.534923 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23f23bd1-c3ee-4ec7-9abb-ca6d2b3d0deb-config\") pod \"route-controller-manager-66dcfbd4c4-7wvzp\" (UID: \"23f23bd1-c3ee-4ec7-9abb-ca6d2b3d0deb\") " pod="openshift-route-controller-manager/route-controller-manager-66dcfbd4c4-7wvzp" Mar 12 00:11:03 crc kubenswrapper[4948]: I0312 00:11:03.534947 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-d9r2j\" (UniqueName: \"kubernetes.io/projected/23f23bd1-c3ee-4ec7-9abb-ca6d2b3d0deb-kube-api-access-d9r2j\") pod \"route-controller-manager-66dcfbd4c4-7wvzp\" (UID: \"23f23bd1-c3ee-4ec7-9abb-ca6d2b3d0deb\") " pod="openshift-route-controller-manager/route-controller-manager-66dcfbd4c4-7wvzp" Mar 12 00:11:03 crc kubenswrapper[4948]: I0312 00:11:03.534987 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25bd67ae-99ad-4f13-a87f-8416e060d3b2-config\") pod \"controller-manager-7f5b4654f4-jnbsn\" (UID: \"25bd67ae-99ad-4f13-a87f-8416e060d3b2\") " pod="openshift-controller-manager/controller-manager-7f5b4654f4-jnbsn" Mar 12 00:11:03 crc kubenswrapper[4948]: I0312 00:11:03.535012 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/23f23bd1-c3ee-4ec7-9abb-ca6d2b3d0deb-client-ca\") pod \"route-controller-manager-66dcfbd4c4-7wvzp\" (UID: \"23f23bd1-c3ee-4ec7-9abb-ca6d2b3d0deb\") " pod="openshift-route-controller-manager/route-controller-manager-66dcfbd4c4-7wvzp" Mar 12 00:11:03 crc kubenswrapper[4948]: I0312 00:11:03.535036 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/25bd67ae-99ad-4f13-a87f-8416e060d3b2-serving-cert\") pod \"controller-manager-7f5b4654f4-jnbsn\" (UID: \"25bd67ae-99ad-4f13-a87f-8416e060d3b2\") " pod="openshift-controller-manager/controller-manager-7f5b4654f4-jnbsn" Mar 12 00:11:03 crc kubenswrapper[4948]: I0312 00:11:03.535069 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/25bd67ae-99ad-4f13-a87f-8416e060d3b2-proxy-ca-bundles\") pod \"controller-manager-7f5b4654f4-jnbsn\" (UID: \"25bd67ae-99ad-4f13-a87f-8416e060d3b2\") " pod="openshift-controller-manager/controller-manager-7f5b4654f4-jnbsn" Mar 12 00:11:03 crc kubenswrapper[4948]: I0312 00:11:03.535135 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/7aa3a362-6bdc-409f-8341-6c0c494f163c-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 12 00:11:03 crc kubenswrapper[4948]: I0312 00:11:03.536000 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23f23bd1-c3ee-4ec7-9abb-ca6d2b3d0deb-config\") pod \"route-controller-manager-66dcfbd4c4-7wvzp\" (UID: \"23f23bd1-c3ee-4ec7-9abb-ca6d2b3d0deb\") " pod="openshift-route-controller-manager/route-controller-manager-66dcfbd4c4-7wvzp" Mar 12 00:11:03 crc kubenswrapper[4948]: I0312 00:11:03.536399 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/23f23bd1-c3ee-4ec7-9abb-ca6d2b3d0deb-client-ca\") pod \"route-controller-manager-66dcfbd4c4-7wvzp\" (UID: \"23f23bd1-c3ee-4ec7-9abb-ca6d2b3d0deb\") " pod="openshift-route-controller-manager/route-controller-manager-66dcfbd4c4-7wvzp" Mar 12 00:11:03 crc kubenswrapper[4948]: I0312 00:11:03.537578 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25bd67ae-99ad-4f13-a87f-8416e060d3b2-config\") pod \"controller-manager-7f5b4654f4-jnbsn\" (UID: \"25bd67ae-99ad-4f13-a87f-8416e060d3b2\") " pod="openshift-controller-manager/controller-manager-7f5b4654f4-jnbsn" Mar 12 00:11:03 crc kubenswrapper[4948]: I0312 00:11:03.537795 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/25bd67ae-99ad-4f13-a87f-8416e060d3b2-proxy-ca-bundles\") pod \"controller-manager-7f5b4654f4-jnbsn\" (UID: \"25bd67ae-99ad-4f13-a87f-8416e060d3b2\") " pod="openshift-controller-manager/controller-manager-7f5b4654f4-jnbsn" Mar 12 00:11:03 crc kubenswrapper[4948]: I0312 00:11:03.542768 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/25bd67ae-99ad-4f13-a87f-8416e060d3b2-serving-cert\") pod \"controller-manager-7f5b4654f4-jnbsn\" (UID: \"25bd67ae-99ad-4f13-a87f-8416e060d3b2\") " pod="openshift-controller-manager/controller-manager-7f5b4654f4-jnbsn" Mar 12 00:11:03 crc kubenswrapper[4948]: I0312 00:11:03.542794 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/23f23bd1-c3ee-4ec7-9abb-ca6d2b3d0deb-serving-cert\") pod \"route-controller-manager-66dcfbd4c4-7wvzp\" (UID: \"23f23bd1-c3ee-4ec7-9abb-ca6d2b3d0deb\") " pod="openshift-route-controller-manager/route-controller-manager-66dcfbd4c4-7wvzp" Mar 12 00:11:03 crc kubenswrapper[4948]: I0312 00:11:03.543991 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/25bd67ae-99ad-4f13-a87f-8416e060d3b2-client-ca\") pod \"controller-manager-7f5b4654f4-jnbsn\" (UID: \"25bd67ae-99ad-4f13-a87f-8416e060d3b2\") " pod="openshift-controller-manager/controller-manager-7f5b4654f4-jnbsn" Mar 12 00:11:03 crc kubenswrapper[4948]: I0312 00:11:03.552217 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-b97hz\" (UniqueName: \"kubernetes.io/projected/25bd67ae-99ad-4f13-a87f-8416e060d3b2-kube-api-access-b97hz\") pod \"controller-manager-7f5b4654f4-jnbsn\" (UID: \"25bd67ae-99ad-4f13-a87f-8416e060d3b2\") " pod="openshift-controller-manager/controller-manager-7f5b4654f4-jnbsn" Mar 12 00:11:03 crc kubenswrapper[4948]: I0312 00:11:03.555029 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-d9r2j\" (UniqueName: \"kubernetes.io/projected/23f23bd1-c3ee-4ec7-9abb-ca6d2b3d0deb-kube-api-access-d9r2j\") pod \"route-controller-manager-66dcfbd4c4-7wvzp\" (UID: \"23f23bd1-c3ee-4ec7-9abb-ca6d2b3d0deb\") " pod="openshift-route-controller-manager/route-controller-manager-66dcfbd4c4-7wvzp" Mar 12 00:11:03 crc kubenswrapper[4948]: I0312 00:11:03.585746 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7f5b4654f4-jnbsn" Mar 12 00:11:03 crc kubenswrapper[4948]: I0312 00:11:03.591906 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-66dcfbd4c4-7wvzp" Mar 12 00:11:03 crc kubenswrapper[4948]: I0312 00:11:03.897405 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-66dcfbd4c4-7wvzp"] Mar 12 00:11:03 crc kubenswrapper[4948]: W0312 00:11:03.908167 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod23f23bd1_c3ee_4ec7_9abb_ca6d2b3d0deb.slice/crio-5c8e0932de21699e210003c636cf194223dcb9522b78b078b934452229261336 WatchSource:0}: Error finding container 5c8e0932de21699e210003c636cf194223dcb9522b78b078b934452229261336: Status 404 returned error can't find the container with id 5c8e0932de21699e210003c636cf194223dcb9522b78b078b934452229261336 Mar 12 00:11:03 crc kubenswrapper[4948]: I0312 00:11:03.948716 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-7f5b4654f4-jnbsn"] Mar 12 00:11:04 crc kubenswrapper[4948]: I0312 00:11:04.037722 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/revision-pruner-9-crc" Mar 12 00:11:04 crc kubenswrapper[4948]: I0312 00:11:04.039374 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/revision-pruner-9-crc" event={"ID":"7aa3a362-6bdc-409f-8341-6c0c494f163c","Type":"ContainerDied","Data":"0e63b2912cc5d71a8361c36427c660ade2286a4d4193a0aef26a13c2aaa5d183"} Mar 12 00:11:04 crc kubenswrapper[4948]: I0312 00:11:04.039418 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="0e63b2912cc5d71a8361c36427c660ade2286a4d4193a0aef26a13c2aaa5d183" Mar 12 00:11:04 crc kubenswrapper[4948]: I0312 00:11:04.049728 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-66dcfbd4c4-7wvzp" event={"ID":"23f23bd1-c3ee-4ec7-9abb-ca6d2b3d0deb","Type":"ContainerStarted","Data":"5c8e0932de21699e210003c636cf194223dcb9522b78b078b934452229261336"} Mar 12 00:11:04 crc kubenswrapper[4948]: I0312 00:11:04.057503 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v8h88" event={"ID":"5d83d568-3968-4ea6-b548-511c951defa5","Type":"ContainerStarted","Data":"e27ca11906dfd90f4412b456fd32a7dce40145941516de818777da1c7d83353c"} Mar 12 00:11:04 crc kubenswrapper[4948]: I0312 00:11:04.059752 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7f5b4654f4-jnbsn" event={"ID":"25bd67ae-99ad-4f13-a87f-8416e060d3b2","Type":"ContainerStarted","Data":"d2ea3333332ca69448406b292cd34c2c9cbd23b39cfa77b04ef926844caabf84"} Mar 12 00:11:04 crc kubenswrapper[4948]: I0312 00:11:04.063485 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29554570-2b2hb" event={"ID":"2fb324d6-44d4-440d-ae60-a5356a3a057f","Type":"ContainerStarted","Data":"6c842526274736c2e138ac271848bc99ec3ab0bb8912406eba8443b61f20cc89"} Mar 12 00:11:04 crc kubenswrapper[4948]: I0312 00:11:04.150240 4948 csr.go:261] certificate signing request csr-psqsz is approved, waiting to be issued Mar 12 00:11:04 crc kubenswrapper[4948]: I0312 00:11:04.157173 4948 csr.go:257] certificate signing request csr-psqsz is issued Mar 12 00:11:05 crc kubenswrapper[4948]: I0312 00:11:05.068154 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7f5b4654f4-jnbsn" event={"ID":"25bd67ae-99ad-4f13-a87f-8416e060d3b2","Type":"ContainerStarted","Data":"c5cdf1ba032993b1192bfec19272b4339c9264b00e52452d109a56007da854f8"} Mar 12 00:11:05 crc kubenswrapper[4948]: I0312 00:11:05.069248 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-7f5b4654f4-jnbsn" Mar 12 00:11:05 crc kubenswrapper[4948]: I0312 00:11:05.070638 4948 generic.go:334] "Generic (PLEG): container finished" podID="2fb324d6-44d4-440d-ae60-a5356a3a057f" containerID="6c842526274736c2e138ac271848bc99ec3ab0bb8912406eba8443b61f20cc89" exitCode=0 Mar 12 00:11:05 crc kubenswrapper[4948]: I0312 00:11:05.070678 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29554570-2b2hb" event={"ID":"2fb324d6-44d4-440d-ae60-a5356a3a057f","Type":"ContainerDied","Data":"6c842526274736c2e138ac271848bc99ec3ab0bb8912406eba8443b61f20cc89"} Mar 12 00:11:05 crc kubenswrapper[4948]: I0312 00:11:05.071646 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-66dcfbd4c4-7wvzp" event={"ID":"23f23bd1-c3ee-4ec7-9abb-ca6d2b3d0deb","Type":"ContainerStarted","Data":"c0099690c0154a146b75cfae69df058dd3e6619ff0ccd594576b65e9783f1aaa"} Mar 12 00:11:05 crc kubenswrapper[4948]: I0312 00:11:05.072073 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-66dcfbd4c4-7wvzp" Mar 12 00:11:05 crc kubenswrapper[4948]: I0312 00:11:05.073428 4948 generic.go:334] "Generic (PLEG): container finished" podID="5d83d568-3968-4ea6-b548-511c951defa5" containerID="e27ca11906dfd90f4412b456fd32a7dce40145941516de818777da1c7d83353c" exitCode=0 Mar 12 00:11:05 crc kubenswrapper[4948]: I0312 00:11:05.073473 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v8h88" event={"ID":"5d83d568-3968-4ea6-b548-511c951defa5","Type":"ContainerDied","Data":"e27ca11906dfd90f4412b456fd32a7dce40145941516de818777da1c7d83353c"} Mar 12 00:11:05 crc kubenswrapper[4948]: I0312 00:11:05.081158 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-66dcfbd4c4-7wvzp" Mar 12 00:11:05 crc kubenswrapper[4948]: I0312 00:11:05.081602 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-7f5b4654f4-jnbsn" Mar 12 00:11:05 crc kubenswrapper[4948]: I0312 00:11:05.092405 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29554570-2b2hb" podStartSLOduration=22.158501712 podStartE2EDuration="1m5.092388984s" podCreationTimestamp="2026-03-12 00:10:00 +0000 UTC" firstStartedPulling="2026-03-12 00:10:20.259431685 +0000 UTC m=+219.715035423" lastFinishedPulling="2026-03-12 00:11:03.193318957 +0000 UTC m=+262.648922695" observedRunningTime="2026-03-12 00:11:04.106418557 +0000 UTC m=+263.562022295" watchObservedRunningTime="2026-03-12 00:11:05.092388984 +0000 UTC m=+264.547992722" Mar 12 00:11:05 crc kubenswrapper[4948]: I0312 00:11:05.130790 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-7f5b4654f4-jnbsn" podStartSLOduration=4.130773472 podStartE2EDuration="4.130773472s" podCreationTimestamp="2026-03-12 00:11:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:11:05.09481663 +0000 UTC m=+264.550420388" watchObservedRunningTime="2026-03-12 00:11:05.130773472 +0000 UTC m=+264.586377220" Mar 12 00:11:05 crc kubenswrapper[4948]: I0312 00:11:05.158086 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-12-13 18:56:51.99917857 +0000 UTC Mar 12 00:11:05 crc kubenswrapper[4948]: I0312 00:11:05.158122 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 6642h45m46.841059625s for next certificate rotation Mar 12 00:11:05 crc kubenswrapper[4948]: I0312 00:11:05.224537 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-66dcfbd4c4-7wvzp" podStartSLOduration=4.224521257 podStartE2EDuration="4.224521257s" podCreationTimestamp="2026-03-12 00:11:01 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:11:05.210043322 +0000 UTC m=+264.665647060" watchObservedRunningTime="2026-03-12 00:11:05.224521257 +0000 UTC m=+264.680124995" Mar 12 00:11:06 crc kubenswrapper[4948]: I0312 00:11:06.158415 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Certificate expiration is 2027-02-24 05:54:36 +0000 UTC, rotation deadline is 2026-12-18 20:26:46.348903109 +0000 UTC Mar 12 00:11:06 crc kubenswrapper[4948]: I0312 00:11:06.158470 4948 certificate_manager.go:356] kubernetes.io/kubelet-serving: Waiting 6764h15m40.190435664s for next certificate rotation Mar 12 00:11:06 crc kubenswrapper[4948]: I0312 00:11:06.331101 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29554570-2b2hb" Mar 12 00:11:06 crc kubenswrapper[4948]: I0312 00:11:06.470612 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6dvsl\" (UniqueName: \"kubernetes.io/projected/2fb324d6-44d4-440d-ae60-a5356a3a057f-kube-api-access-6dvsl\") pod \"2fb324d6-44d4-440d-ae60-a5356a3a057f\" (UID: \"2fb324d6-44d4-440d-ae60-a5356a3a057f\") " Mar 12 00:11:06 crc kubenswrapper[4948]: I0312 00:11:06.478006 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2fb324d6-44d4-440d-ae60-a5356a3a057f-kube-api-access-6dvsl" (OuterVolumeSpecName: "kube-api-access-6dvsl") pod "2fb324d6-44d4-440d-ae60-a5356a3a057f" (UID: "2fb324d6-44d4-440d-ae60-a5356a3a057f"). InnerVolumeSpecName "kube-api-access-6dvsl". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:11:06 crc kubenswrapper[4948]: I0312 00:11:06.572031 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6dvsl\" (UniqueName: \"kubernetes.io/projected/2fb324d6-44d4-440d-ae60-a5356a3a057f-kube-api-access-6dvsl\") on node \"crc\" DevicePath \"\"" Mar 12 00:11:07 crc kubenswrapper[4948]: I0312 00:11:07.091818 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29554570-2b2hb" event={"ID":"2fb324d6-44d4-440d-ae60-a5356a3a057f","Type":"ContainerDied","Data":"1d4d414e0f4148242881bbd99b11885b2652663b62baaf55ce4bb47096bd3a56"} Mar 12 00:11:07 crc kubenswrapper[4948]: I0312 00:11:07.092421 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="1d4d414e0f4148242881bbd99b11885b2652663b62baaf55ce4bb47096bd3a56" Mar 12 00:11:07 crc kubenswrapper[4948]: I0312 00:11:07.092343 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29554570-2b2hb" Mar 12 00:11:07 crc kubenswrapper[4948]: I0312 00:11:07.162366 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-zc9zh" Mar 12 00:11:07 crc kubenswrapper[4948]: I0312 00:11:07.162544 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-zc9zh" Mar 12 00:11:07 crc kubenswrapper[4948]: I0312 00:11:07.574219 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-bsmcp" Mar 12 00:11:07 crc kubenswrapper[4948]: I0312 00:11:07.574353 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-bsmcp" Mar 12 00:11:07 crc kubenswrapper[4948]: I0312 00:11:07.779126 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-bsmcp" Mar 12 00:11:07 crc kubenswrapper[4948]: I0312 00:11:07.780431 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-zc9zh" Mar 12 00:11:07 crc kubenswrapper[4948]: I0312 00:11:07.894597 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 12 00:11:07 crc kubenswrapper[4948]: E0312 00:11:07.894830 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2fb324d6-44d4-440d-ae60-a5356a3a057f" containerName="oc" Mar 12 00:11:07 crc kubenswrapper[4948]: I0312 00:11:07.894844 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="2fb324d6-44d4-440d-ae60-a5356a3a057f" containerName="oc" Mar 12 00:11:07 crc kubenswrapper[4948]: E0312 00:11:07.894855 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7aa3a362-6bdc-409f-8341-6c0c494f163c" containerName="pruner" Mar 12 00:11:07 crc kubenswrapper[4948]: I0312 00:11:07.894861 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="7aa3a362-6bdc-409f-8341-6c0c494f163c" containerName="pruner" Mar 12 00:11:07 crc kubenswrapper[4948]: I0312 00:11:07.895397 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="7aa3a362-6bdc-409f-8341-6c0c494f163c" containerName="pruner" Mar 12 00:11:07 crc kubenswrapper[4948]: I0312 00:11:07.895423 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="2fb324d6-44d4-440d-ae60-a5356a3a057f" containerName="oc" Mar 12 00:11:07 crc kubenswrapper[4948]: I0312 00:11:07.899024 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 12 00:11:07 crc kubenswrapper[4948]: I0312 00:11:07.909643 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver"/"kube-root-ca.crt" Mar 12 00:11:07 crc kubenswrapper[4948]: I0312 00:11:07.910536 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver"/"installer-sa-dockercfg-5pr6n" Mar 12 00:11:07 crc kubenswrapper[4948]: I0312 00:11:07.917976 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 12 00:11:08 crc kubenswrapper[4948]: I0312 00:11:08.093046 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/62c7736b-4d37-43bf-8ee8-f1e5c3ac3a37-kube-api-access\") pod \"installer-9-crc\" (UID: \"62c7736b-4d37-43bf-8ee8-f1e5c3ac3a37\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 12 00:11:08 crc kubenswrapper[4948]: I0312 00:11:08.093624 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/62c7736b-4d37-43bf-8ee8-f1e5c3ac3a37-var-lock\") pod \"installer-9-crc\" (UID: \"62c7736b-4d37-43bf-8ee8-f1e5c3ac3a37\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 12 00:11:08 crc kubenswrapper[4948]: I0312 00:11:08.093777 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/62c7736b-4d37-43bf-8ee8-f1e5c3ac3a37-kubelet-dir\") pod \"installer-9-crc\" (UID: \"62c7736b-4d37-43bf-8ee8-f1e5c3ac3a37\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 12 00:11:08 crc kubenswrapper[4948]: I0312 00:11:08.099053 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v8h88" event={"ID":"5d83d568-3968-4ea6-b548-511c951defa5","Type":"ContainerStarted","Data":"05fd166aeeb70277dab1df229c8cbde7b85139a276d0732903b8b28713344011"} Mar 12 00:11:08 crc kubenswrapper[4948]: I0312 00:11:08.121910 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-v8h88" podStartSLOduration=3.958467611 podStartE2EDuration="44.121886914s" podCreationTimestamp="2026-03-12 00:10:24 +0000 UTC" firstStartedPulling="2026-03-12 00:10:26.53569719 +0000 UTC m=+225.991300928" lastFinishedPulling="2026-03-12 00:11:06.699116493 +0000 UTC m=+266.154720231" observedRunningTime="2026-03-12 00:11:08.117891554 +0000 UTC m=+267.573495292" watchObservedRunningTime="2026-03-12 00:11:08.121886914 +0000 UTC m=+267.577490662" Mar 12 00:11:08 crc kubenswrapper[4948]: I0312 00:11:08.153057 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-zc9zh" Mar 12 00:11:08 crc kubenswrapper[4948]: I0312 00:11:08.162925 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-bsmcp" Mar 12 00:11:08 crc kubenswrapper[4948]: I0312 00:11:08.194596 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/62c7736b-4d37-43bf-8ee8-f1e5c3ac3a37-var-lock\") pod \"installer-9-crc\" (UID: \"62c7736b-4d37-43bf-8ee8-f1e5c3ac3a37\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 12 00:11:08 crc kubenswrapper[4948]: I0312 00:11:08.194678 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/62c7736b-4d37-43bf-8ee8-f1e5c3ac3a37-kubelet-dir\") pod \"installer-9-crc\" (UID: \"62c7736b-4d37-43bf-8ee8-f1e5c3ac3a37\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 12 00:11:08 crc kubenswrapper[4948]: I0312 00:11:08.194713 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/62c7736b-4d37-43bf-8ee8-f1e5c3ac3a37-kube-api-access\") pod \"installer-9-crc\" (UID: \"62c7736b-4d37-43bf-8ee8-f1e5c3ac3a37\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 12 00:11:08 crc kubenswrapper[4948]: I0312 00:11:08.195130 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/62c7736b-4d37-43bf-8ee8-f1e5c3ac3a37-var-lock\") pod \"installer-9-crc\" (UID: \"62c7736b-4d37-43bf-8ee8-f1e5c3ac3a37\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 12 00:11:08 crc kubenswrapper[4948]: I0312 00:11:08.195604 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/62c7736b-4d37-43bf-8ee8-f1e5c3ac3a37-kubelet-dir\") pod \"installer-9-crc\" (UID: \"62c7736b-4d37-43bf-8ee8-f1e5c3ac3a37\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 12 00:11:08 crc kubenswrapper[4948]: I0312 00:11:08.214087 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-btd4j" Mar 12 00:11:08 crc kubenswrapper[4948]: I0312 00:11:08.214127 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-btd4j" Mar 12 00:11:08 crc kubenswrapper[4948]: I0312 00:11:08.220407 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/62c7736b-4d37-43bf-8ee8-f1e5c3ac3a37-kube-api-access\") pod \"installer-9-crc\" (UID: \"62c7736b-4d37-43bf-8ee8-f1e5c3ac3a37\") " pod="openshift-kube-apiserver/installer-9-crc" Mar 12 00:11:08 crc kubenswrapper[4948]: I0312 00:11:08.222237 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 12 00:11:08 crc kubenswrapper[4948]: I0312 00:11:08.255550 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-btd4j" Mar 12 00:11:08 crc kubenswrapper[4948]: I0312 00:11:08.374907 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-6w6fq" Mar 12 00:11:08 crc kubenswrapper[4948]: I0312 00:11:08.375289 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-6w6fq" Mar 12 00:11:08 crc kubenswrapper[4948]: I0312 00:11:08.422684 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-6w6fq" Mar 12 00:11:08 crc kubenswrapper[4948]: I0312 00:11:08.663025 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-kube-apiserver/installer-9-crc"] Mar 12 00:11:09 crc kubenswrapper[4948]: I0312 00:11:09.106063 4948 generic.go:334] "Generic (PLEG): container finished" podID="239cfe46-3a3a-4624-9aa2-ea7112f28947" containerID="1773b11c62a5767b2972ce020f88a29bd0e48a9ebb0cad24731ba22620298d87" exitCode=0 Mar 12 00:11:09 crc kubenswrapper[4948]: I0312 00:11:09.106486 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hcvm5" event={"ID":"239cfe46-3a3a-4624-9aa2-ea7112f28947","Type":"ContainerDied","Data":"1773b11c62a5767b2972ce020f88a29bd0e48a9ebb0cad24731ba22620298d87"} Mar 12 00:11:09 crc kubenswrapper[4948]: I0312 00:11:09.108943 4948 generic.go:334] "Generic (PLEG): container finished" podID="e6318b61-ef2d-41da-8c92-ff7523ce613f" containerID="8848e7cb0bd5b2eee15cd54c7a8a9b39bffb36f27a9fc09e35e7884845d59b5b" exitCode=0 Mar 12 00:11:09 crc kubenswrapper[4948]: I0312 00:11:09.109398 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-62wnf" event={"ID":"e6318b61-ef2d-41da-8c92-ff7523ce613f","Type":"ContainerDied","Data":"8848e7cb0bd5b2eee15cd54c7a8a9b39bffb36f27a9fc09e35e7884845d59b5b"} Mar 12 00:11:09 crc kubenswrapper[4948]: I0312 00:11:09.112574 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"62c7736b-4d37-43bf-8ee8-f1e5c3ac3a37","Type":"ContainerStarted","Data":"9b56865be8299fc5e275b1124ac672af03fb948b0c3ddddb03d781cef1a45df3"} Mar 12 00:11:09 crc kubenswrapper[4948]: I0312 00:11:09.172866 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-btd4j" Mar 12 00:11:09 crc kubenswrapper[4948]: I0312 00:11:09.175386 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-6w6fq" Mar 12 00:11:10 crc kubenswrapper[4948]: I0312 00:11:10.126897 4948 generic.go:334] "Generic (PLEG): container finished" podID="5184125a-dde1-4db4-9d6b-5b3c45d8d82a" containerID="34b02d06deb5bf3f40bbae42cf819536fab8af8ed2c6cc5d9120f3ba633efc66" exitCode=0 Mar 12 00:11:10 crc kubenswrapper[4948]: I0312 00:11:10.126963 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rflhd" event={"ID":"5184125a-dde1-4db4-9d6b-5b3c45d8d82a","Type":"ContainerDied","Data":"34b02d06deb5bf3f40bbae42cf819536fab8af8ed2c6cc5d9120f3ba633efc66"} Mar 12 00:11:10 crc kubenswrapper[4948]: I0312 00:11:10.130818 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"62c7736b-4d37-43bf-8ee8-f1e5c3ac3a37","Type":"ContainerStarted","Data":"134eb20d8890c76b8433a37aaa725ef8f2322110a3d18539d34370335a2841f3"} Mar 12 00:11:10 crc kubenswrapper[4948]: I0312 00:11:10.135245 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hcvm5" event={"ID":"239cfe46-3a3a-4624-9aa2-ea7112f28947","Type":"ContainerStarted","Data":"6c529a18c3f1f58c2f5de6e8fd7069a26d1d10de15f10b482d95979542b13de1"} Mar 12 00:11:10 crc kubenswrapper[4948]: I0312 00:11:10.137814 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-62wnf" event={"ID":"e6318b61-ef2d-41da-8c92-ff7523ce613f","Type":"ContainerStarted","Data":"663949e58f30470b63459dc123ff8a880fc6ba09cf46402c77d36e17f4cab6fb"} Mar 12 00:11:10 crc kubenswrapper[4948]: I0312 00:11:10.160126 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-62wnf" podStartSLOduration=1.9586109619999998 podStartE2EDuration="45.160109428s" podCreationTimestamp="2026-03-12 00:10:25 +0000 UTC" firstStartedPulling="2026-03-12 00:10:26.538414883 +0000 UTC m=+225.994018621" lastFinishedPulling="2026-03-12 00:11:09.739913339 +0000 UTC m=+269.195517087" observedRunningTime="2026-03-12 00:11:10.159096272 +0000 UTC m=+269.614700010" watchObservedRunningTime="2026-03-12 00:11:10.160109428 +0000 UTC m=+269.615713166" Mar 12 00:11:10 crc kubenswrapper[4948]: I0312 00:11:10.176359 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-hcvm5" podStartSLOduration=2.03981474 podStartE2EDuration="45.176322536s" podCreationTimestamp="2026-03-12 00:10:25 +0000 UTC" firstStartedPulling="2026-03-12 00:10:26.533853066 +0000 UTC m=+225.989456804" lastFinishedPulling="2026-03-12 00:11:09.670360852 +0000 UTC m=+269.125964600" observedRunningTime="2026-03-12 00:11:10.174046733 +0000 UTC m=+269.629650481" watchObservedRunningTime="2026-03-12 00:11:10.176322536 +0000 UTC m=+269.631926294" Mar 12 00:11:11 crc kubenswrapper[4948]: I0312 00:11:11.144710 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rflhd" event={"ID":"5184125a-dde1-4db4-9d6b-5b3c45d8d82a","Type":"ContainerStarted","Data":"bf689a8e9a12667b22bba99be51be88469d4e2cd090bfa6efe2c69a940fe8322"} Mar 12 00:11:11 crc kubenswrapper[4948]: I0312 00:11:11.164968 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-rflhd" podStartSLOduration=3.150612478 podStartE2EDuration="47.164945093s" podCreationTimestamp="2026-03-12 00:10:24 +0000 UTC" firstStartedPulling="2026-03-12 00:10:26.541994066 +0000 UTC m=+225.997597804" lastFinishedPulling="2026-03-12 00:11:10.556326681 +0000 UTC m=+270.011930419" observedRunningTime="2026-03-12 00:11:11.164520107 +0000 UTC m=+270.620123845" watchObservedRunningTime="2026-03-12 00:11:11.164945093 +0000 UTC m=+270.620548851" Mar 12 00:11:11 crc kubenswrapper[4948]: I0312 00:11:11.166579 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/installer-9-crc" podStartSLOduration=4.166572068 podStartE2EDuration="4.166572068s" podCreationTimestamp="2026-03-12 00:11:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:11:10.191044644 +0000 UTC m=+269.646648382" watchObservedRunningTime="2026-03-12 00:11:11.166572068 +0000 UTC m=+270.622175826" Mar 12 00:11:11 crc kubenswrapper[4948]: I0312 00:11:11.238268 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bsmcp"] Mar 12 00:11:11 crc kubenswrapper[4948]: I0312 00:11:11.238540 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-bsmcp" podUID="44212e8e-c675-4643-a752-997be5f8e980" containerName="registry-server" containerID="cri-o://d012b13a227c5089a7a52d6cf33547053badf64ce2fbf7f65be8dec74ef2954f" gracePeriod=2 Mar 12 00:11:11 crc kubenswrapper[4948]: I0312 00:11:11.439136 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6w6fq"] Mar 12 00:11:11 crc kubenswrapper[4948]: I0312 00:11:11.439401 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-6w6fq" podUID="529630b8-cc23-44b7-bafd-6ae1313877d1" containerName="registry-server" containerID="cri-o://a40fecd469f3d505800497aa8b7497e5f3c26dbab5363fa4708bc6ad58b76789" gracePeriod=2 Mar 12 00:11:11 crc kubenswrapper[4948]: I0312 00:11:11.750455 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bsmcp" Mar 12 00:11:11 crc kubenswrapper[4948]: I0312 00:11:11.854749 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44212e8e-c675-4643-a752-997be5f8e980-catalog-content\") pod \"44212e8e-c675-4643-a752-997be5f8e980\" (UID: \"44212e8e-c675-4643-a752-997be5f8e980\") " Mar 12 00:11:11 crc kubenswrapper[4948]: I0312 00:11:11.854849 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-h267n\" (UniqueName: \"kubernetes.io/projected/44212e8e-c675-4643-a752-997be5f8e980-kube-api-access-h267n\") pod \"44212e8e-c675-4643-a752-997be5f8e980\" (UID: \"44212e8e-c675-4643-a752-997be5f8e980\") " Mar 12 00:11:11 crc kubenswrapper[4948]: I0312 00:11:11.854907 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44212e8e-c675-4643-a752-997be5f8e980-utilities\") pod \"44212e8e-c675-4643-a752-997be5f8e980\" (UID: \"44212e8e-c675-4643-a752-997be5f8e980\") " Mar 12 00:11:11 crc kubenswrapper[4948]: I0312 00:11:11.856350 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44212e8e-c675-4643-a752-997be5f8e980-utilities" (OuterVolumeSpecName: "utilities") pod "44212e8e-c675-4643-a752-997be5f8e980" (UID: "44212e8e-c675-4643-a752-997be5f8e980"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 00:11:11 crc kubenswrapper[4948]: I0312 00:11:11.860509 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/44212e8e-c675-4643-a752-997be5f8e980-kube-api-access-h267n" (OuterVolumeSpecName: "kube-api-access-h267n") pod "44212e8e-c675-4643-a752-997be5f8e980" (UID: "44212e8e-c675-4643-a752-997be5f8e980"). InnerVolumeSpecName "kube-api-access-h267n". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:11:11 crc kubenswrapper[4948]: I0312 00:11:11.880814 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6w6fq" Mar 12 00:11:11 crc kubenswrapper[4948]: I0312 00:11:11.895245 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/44212e8e-c675-4643-a752-997be5f8e980-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "44212e8e-c675-4643-a752-997be5f8e980" (UID: "44212e8e-c675-4643-a752-997be5f8e980"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 00:11:11 crc kubenswrapper[4948]: I0312 00:11:11.956044 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/44212e8e-c675-4643-a752-997be5f8e980-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 00:11:11 crc kubenswrapper[4948]: I0312 00:11:11.956081 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-h267n\" (UniqueName: \"kubernetes.io/projected/44212e8e-c675-4643-a752-997be5f8e980-kube-api-access-h267n\") on node \"crc\" DevicePath \"\"" Mar 12 00:11:11 crc kubenswrapper[4948]: I0312 00:11:11.956109 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/44212e8e-c675-4643-a752-997be5f8e980-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 00:11:12 crc kubenswrapper[4948]: I0312 00:11:12.056829 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/529630b8-cc23-44b7-bafd-6ae1313877d1-utilities\") pod \"529630b8-cc23-44b7-bafd-6ae1313877d1\" (UID: \"529630b8-cc23-44b7-bafd-6ae1313877d1\") " Mar 12 00:11:12 crc kubenswrapper[4948]: I0312 00:11:12.056912 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/529630b8-cc23-44b7-bafd-6ae1313877d1-catalog-content\") pod \"529630b8-cc23-44b7-bafd-6ae1313877d1\" (UID: \"529630b8-cc23-44b7-bafd-6ae1313877d1\") " Mar 12 00:11:12 crc kubenswrapper[4948]: I0312 00:11:12.056948 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-zx4xz\" (UniqueName: \"kubernetes.io/projected/529630b8-cc23-44b7-bafd-6ae1313877d1-kube-api-access-zx4xz\") pod \"529630b8-cc23-44b7-bafd-6ae1313877d1\" (UID: \"529630b8-cc23-44b7-bafd-6ae1313877d1\") " Mar 12 00:11:12 crc kubenswrapper[4948]: I0312 00:11:12.057707 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/529630b8-cc23-44b7-bafd-6ae1313877d1-utilities" (OuterVolumeSpecName: "utilities") pod "529630b8-cc23-44b7-bafd-6ae1313877d1" (UID: "529630b8-cc23-44b7-bafd-6ae1313877d1"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 00:11:12 crc kubenswrapper[4948]: I0312 00:11:12.061450 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/529630b8-cc23-44b7-bafd-6ae1313877d1-kube-api-access-zx4xz" (OuterVolumeSpecName: "kube-api-access-zx4xz") pod "529630b8-cc23-44b7-bafd-6ae1313877d1" (UID: "529630b8-cc23-44b7-bafd-6ae1313877d1"). InnerVolumeSpecName "kube-api-access-zx4xz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:11:12 crc kubenswrapper[4948]: I0312 00:11:12.156791 4948 generic.go:334] "Generic (PLEG): container finished" podID="529630b8-cc23-44b7-bafd-6ae1313877d1" containerID="a40fecd469f3d505800497aa8b7497e5f3c26dbab5363fa4708bc6ad58b76789" exitCode=0 Mar 12 00:11:12 crc kubenswrapper[4948]: I0312 00:11:12.156867 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6w6fq" event={"ID":"529630b8-cc23-44b7-bafd-6ae1313877d1","Type":"ContainerDied","Data":"a40fecd469f3d505800497aa8b7497e5f3c26dbab5363fa4708bc6ad58b76789"} Mar 12 00:11:12 crc kubenswrapper[4948]: I0312 00:11:12.156898 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-6w6fq" event={"ID":"529630b8-cc23-44b7-bafd-6ae1313877d1","Type":"ContainerDied","Data":"f035a5c066dcd602a7c0553774a03c5978745c83f0997e160378b6f3b4c716d0"} Mar 12 00:11:12 crc kubenswrapper[4948]: I0312 00:11:12.156919 4948 scope.go:117] "RemoveContainer" containerID="a40fecd469f3d505800497aa8b7497e5f3c26dbab5363fa4708bc6ad58b76789" Mar 12 00:11:12 crc kubenswrapper[4948]: I0312 00:11:12.156915 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-6w6fq" Mar 12 00:11:12 crc kubenswrapper[4948]: I0312 00:11:12.157826 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/529630b8-cc23-44b7-bafd-6ae1313877d1-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 00:11:12 crc kubenswrapper[4948]: I0312 00:11:12.157851 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-zx4xz\" (UniqueName: \"kubernetes.io/projected/529630b8-cc23-44b7-bafd-6ae1313877d1-kube-api-access-zx4xz\") on node \"crc\" DevicePath \"\"" Mar 12 00:11:12 crc kubenswrapper[4948]: I0312 00:11:12.163963 4948 generic.go:334] "Generic (PLEG): container finished" podID="44212e8e-c675-4643-a752-997be5f8e980" containerID="d012b13a227c5089a7a52d6cf33547053badf64ce2fbf7f65be8dec74ef2954f" exitCode=0 Mar 12 00:11:12 crc kubenswrapper[4948]: I0312 00:11:12.164039 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-bsmcp" Mar 12 00:11:12 crc kubenswrapper[4948]: I0312 00:11:12.164040 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bsmcp" event={"ID":"44212e8e-c675-4643-a752-997be5f8e980","Type":"ContainerDied","Data":"d012b13a227c5089a7a52d6cf33547053badf64ce2fbf7f65be8dec74ef2954f"} Mar 12 00:11:12 crc kubenswrapper[4948]: I0312 00:11:12.164123 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-bsmcp" event={"ID":"44212e8e-c675-4643-a752-997be5f8e980","Type":"ContainerDied","Data":"05d255416a0db402af61232adca03cefca9928f3acf1a396523501205728b4a8"} Mar 12 00:11:12 crc kubenswrapper[4948]: I0312 00:11:12.189198 4948 scope.go:117] "RemoveContainer" containerID="5ed67fc29cf834f1e10fc0bf2dfcb81ae013ba954f4ebeaf9e1f55017ff9e715" Mar 12 00:11:12 crc kubenswrapper[4948]: I0312 00:11:12.200164 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-bsmcp"] Mar 12 00:11:12 crc kubenswrapper[4948]: I0312 00:11:12.202933 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-bsmcp"] Mar 12 00:11:12 crc kubenswrapper[4948]: I0312 00:11:12.208965 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/529630b8-cc23-44b7-bafd-6ae1313877d1-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "529630b8-cc23-44b7-bafd-6ae1313877d1" (UID: "529630b8-cc23-44b7-bafd-6ae1313877d1"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 00:11:12 crc kubenswrapper[4948]: I0312 00:11:12.225482 4948 scope.go:117] "RemoveContainer" containerID="e18323d8a5d040179826f150b6099523db491ee1abb8b59ec1cb4aa4f795efdd" Mar 12 00:11:12 crc kubenswrapper[4948]: I0312 00:11:12.249686 4948 scope.go:117] "RemoveContainer" containerID="a40fecd469f3d505800497aa8b7497e5f3c26dbab5363fa4708bc6ad58b76789" Mar 12 00:11:12 crc kubenswrapper[4948]: E0312 00:11:12.250375 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a40fecd469f3d505800497aa8b7497e5f3c26dbab5363fa4708bc6ad58b76789\": container with ID starting with a40fecd469f3d505800497aa8b7497e5f3c26dbab5363fa4708bc6ad58b76789 not found: ID does not exist" containerID="a40fecd469f3d505800497aa8b7497e5f3c26dbab5363fa4708bc6ad58b76789" Mar 12 00:11:12 crc kubenswrapper[4948]: I0312 00:11:12.250434 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a40fecd469f3d505800497aa8b7497e5f3c26dbab5363fa4708bc6ad58b76789"} err="failed to get container status \"a40fecd469f3d505800497aa8b7497e5f3c26dbab5363fa4708bc6ad58b76789\": rpc error: code = NotFound desc = could not find container \"a40fecd469f3d505800497aa8b7497e5f3c26dbab5363fa4708bc6ad58b76789\": container with ID starting with a40fecd469f3d505800497aa8b7497e5f3c26dbab5363fa4708bc6ad58b76789 not found: ID does not exist" Mar 12 00:11:12 crc kubenswrapper[4948]: I0312 00:11:12.250460 4948 scope.go:117] "RemoveContainer" containerID="5ed67fc29cf834f1e10fc0bf2dfcb81ae013ba954f4ebeaf9e1f55017ff9e715" Mar 12 00:11:12 crc kubenswrapper[4948]: E0312 00:11:12.250997 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"5ed67fc29cf834f1e10fc0bf2dfcb81ae013ba954f4ebeaf9e1f55017ff9e715\": container with ID starting with 5ed67fc29cf834f1e10fc0bf2dfcb81ae013ba954f4ebeaf9e1f55017ff9e715 not found: ID does not exist" containerID="5ed67fc29cf834f1e10fc0bf2dfcb81ae013ba954f4ebeaf9e1f55017ff9e715" Mar 12 00:11:12 crc kubenswrapper[4948]: I0312 00:11:12.251020 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"5ed67fc29cf834f1e10fc0bf2dfcb81ae013ba954f4ebeaf9e1f55017ff9e715"} err="failed to get container status \"5ed67fc29cf834f1e10fc0bf2dfcb81ae013ba954f4ebeaf9e1f55017ff9e715\": rpc error: code = NotFound desc = could not find container \"5ed67fc29cf834f1e10fc0bf2dfcb81ae013ba954f4ebeaf9e1f55017ff9e715\": container with ID starting with 5ed67fc29cf834f1e10fc0bf2dfcb81ae013ba954f4ebeaf9e1f55017ff9e715 not found: ID does not exist" Mar 12 00:11:12 crc kubenswrapper[4948]: I0312 00:11:12.251038 4948 scope.go:117] "RemoveContainer" containerID="e18323d8a5d040179826f150b6099523db491ee1abb8b59ec1cb4aa4f795efdd" Mar 12 00:11:12 crc kubenswrapper[4948]: E0312 00:11:12.251489 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"e18323d8a5d040179826f150b6099523db491ee1abb8b59ec1cb4aa4f795efdd\": container with ID starting with e18323d8a5d040179826f150b6099523db491ee1abb8b59ec1cb4aa4f795efdd not found: ID does not exist" containerID="e18323d8a5d040179826f150b6099523db491ee1abb8b59ec1cb4aa4f795efdd" Mar 12 00:11:12 crc kubenswrapper[4948]: I0312 00:11:12.251539 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"e18323d8a5d040179826f150b6099523db491ee1abb8b59ec1cb4aa4f795efdd"} err="failed to get container status \"e18323d8a5d040179826f150b6099523db491ee1abb8b59ec1cb4aa4f795efdd\": rpc error: code = NotFound desc = could not find container \"e18323d8a5d040179826f150b6099523db491ee1abb8b59ec1cb4aa4f795efdd\": container with ID starting with e18323d8a5d040179826f150b6099523db491ee1abb8b59ec1cb4aa4f795efdd not found: ID does not exist" Mar 12 00:11:12 crc kubenswrapper[4948]: I0312 00:11:12.251558 4948 scope.go:117] "RemoveContainer" containerID="d012b13a227c5089a7a52d6cf33547053badf64ce2fbf7f65be8dec74ef2954f" Mar 12 00:11:12 crc kubenswrapper[4948]: I0312 00:11:12.259588 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/529630b8-cc23-44b7-bafd-6ae1313877d1-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 00:11:12 crc kubenswrapper[4948]: I0312 00:11:12.271506 4948 scope.go:117] "RemoveContainer" containerID="b87f1313ce649ed1dd52e80afc20736b50c50e9bf8e2e84c9837a9177cd604ba" Mar 12 00:11:12 crc kubenswrapper[4948]: I0312 00:11:12.299885 4948 scope.go:117] "RemoveContainer" containerID="62ce39169fa39ff6e1b56e3fea7f604f7e9f4fd54c2e527988836d1a811b10e3" Mar 12 00:11:12 crc kubenswrapper[4948]: I0312 00:11:12.329058 4948 scope.go:117] "RemoveContainer" containerID="d012b13a227c5089a7a52d6cf33547053badf64ce2fbf7f65be8dec74ef2954f" Mar 12 00:11:12 crc kubenswrapper[4948]: E0312 00:11:12.329502 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d012b13a227c5089a7a52d6cf33547053badf64ce2fbf7f65be8dec74ef2954f\": container with ID starting with d012b13a227c5089a7a52d6cf33547053badf64ce2fbf7f65be8dec74ef2954f not found: ID does not exist" containerID="d012b13a227c5089a7a52d6cf33547053badf64ce2fbf7f65be8dec74ef2954f" Mar 12 00:11:12 crc kubenswrapper[4948]: I0312 00:11:12.329532 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d012b13a227c5089a7a52d6cf33547053badf64ce2fbf7f65be8dec74ef2954f"} err="failed to get container status \"d012b13a227c5089a7a52d6cf33547053badf64ce2fbf7f65be8dec74ef2954f\": rpc error: code = NotFound desc = could not find container \"d012b13a227c5089a7a52d6cf33547053badf64ce2fbf7f65be8dec74ef2954f\": container with ID starting with d012b13a227c5089a7a52d6cf33547053badf64ce2fbf7f65be8dec74ef2954f not found: ID does not exist" Mar 12 00:11:12 crc kubenswrapper[4948]: I0312 00:11:12.329559 4948 scope.go:117] "RemoveContainer" containerID="b87f1313ce649ed1dd52e80afc20736b50c50e9bf8e2e84c9837a9177cd604ba" Mar 12 00:11:12 crc kubenswrapper[4948]: E0312 00:11:12.331874 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b87f1313ce649ed1dd52e80afc20736b50c50e9bf8e2e84c9837a9177cd604ba\": container with ID starting with b87f1313ce649ed1dd52e80afc20736b50c50e9bf8e2e84c9837a9177cd604ba not found: ID does not exist" containerID="b87f1313ce649ed1dd52e80afc20736b50c50e9bf8e2e84c9837a9177cd604ba" Mar 12 00:11:12 crc kubenswrapper[4948]: I0312 00:11:12.331901 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b87f1313ce649ed1dd52e80afc20736b50c50e9bf8e2e84c9837a9177cd604ba"} err="failed to get container status \"b87f1313ce649ed1dd52e80afc20736b50c50e9bf8e2e84c9837a9177cd604ba\": rpc error: code = NotFound desc = could not find container \"b87f1313ce649ed1dd52e80afc20736b50c50e9bf8e2e84c9837a9177cd604ba\": container with ID starting with b87f1313ce649ed1dd52e80afc20736b50c50e9bf8e2e84c9837a9177cd604ba not found: ID does not exist" Mar 12 00:11:12 crc kubenswrapper[4948]: I0312 00:11:12.331915 4948 scope.go:117] "RemoveContainer" containerID="62ce39169fa39ff6e1b56e3fea7f604f7e9f4fd54c2e527988836d1a811b10e3" Mar 12 00:11:12 crc kubenswrapper[4948]: E0312 00:11:12.333098 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"62ce39169fa39ff6e1b56e3fea7f604f7e9f4fd54c2e527988836d1a811b10e3\": container with ID starting with 62ce39169fa39ff6e1b56e3fea7f604f7e9f4fd54c2e527988836d1a811b10e3 not found: ID does not exist" containerID="62ce39169fa39ff6e1b56e3fea7f604f7e9f4fd54c2e527988836d1a811b10e3" Mar 12 00:11:12 crc kubenswrapper[4948]: I0312 00:11:12.333176 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"62ce39169fa39ff6e1b56e3fea7f604f7e9f4fd54c2e527988836d1a811b10e3"} err="failed to get container status \"62ce39169fa39ff6e1b56e3fea7f604f7e9f4fd54c2e527988836d1a811b10e3\": rpc error: code = NotFound desc = could not find container \"62ce39169fa39ff6e1b56e3fea7f604f7e9f4fd54c2e527988836d1a811b10e3\": container with ID starting with 62ce39169fa39ff6e1b56e3fea7f604f7e9f4fd54c2e527988836d1a811b10e3 not found: ID does not exist" Mar 12 00:11:12 crc kubenswrapper[4948]: I0312 00:11:12.489791 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-6w6fq"] Mar 12 00:11:12 crc kubenswrapper[4948]: I0312 00:11:12.494067 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-6w6fq"] Mar 12 00:11:13 crc kubenswrapper[4948]: I0312 00:11:13.324289 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="44212e8e-c675-4643-a752-997be5f8e980" path="/var/lib/kubelet/pods/44212e8e-c675-4643-a752-997be5f8e980/volumes" Mar 12 00:11:13 crc kubenswrapper[4948]: I0312 00:11:13.326282 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="529630b8-cc23-44b7-bafd-6ae1313877d1" path="/var/lib/kubelet/pods/529630b8-cc23-44b7-bafd-6ae1313877d1/volumes" Mar 12 00:11:14 crc kubenswrapper[4948]: I0312 00:11:14.785265 4948 patch_prober.go:28] interesting pod/machine-config-daemon-m4xwc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 00:11:14 crc kubenswrapper[4948]: I0312 00:11:14.785639 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" podUID="7e483c2b-08f0-4e92-8e4a-b7281f30af3e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 00:11:14 crc kubenswrapper[4948]: I0312 00:11:14.785707 4948 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" Mar 12 00:11:14 crc kubenswrapper[4948]: I0312 00:11:14.786551 4948 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"82de8f2b633f1b6c295149d077bfd2b308f5acaa429acf411ed16f456536d0ba"} pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 12 00:11:14 crc kubenswrapper[4948]: I0312 00:11:14.786680 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" podUID="7e483c2b-08f0-4e92-8e4a-b7281f30af3e" containerName="machine-config-daemon" containerID="cri-o://82de8f2b633f1b6c295149d077bfd2b308f5acaa429acf411ed16f456536d0ba" gracePeriod=600 Mar 12 00:11:14 crc kubenswrapper[4948]: I0312 00:11:14.962112 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-v8h88" Mar 12 00:11:14 crc kubenswrapper[4948]: I0312 00:11:14.962156 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-v8h88" Mar 12 00:11:15 crc kubenswrapper[4948]: I0312 00:11:15.017257 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-v8h88" Mar 12 00:11:15 crc kubenswrapper[4948]: I0312 00:11:15.163411 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-rflhd" Mar 12 00:11:15 crc kubenswrapper[4948]: I0312 00:11:15.163478 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-rflhd" Mar 12 00:11:15 crc kubenswrapper[4948]: I0312 00:11:15.185374 4948 generic.go:334] "Generic (PLEG): container finished" podID="7e483c2b-08f0-4e92-8e4a-b7281f30af3e" containerID="82de8f2b633f1b6c295149d077bfd2b308f5acaa429acf411ed16f456536d0ba" exitCode=0 Mar 12 00:11:15 crc kubenswrapper[4948]: I0312 00:11:15.185494 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" event={"ID":"7e483c2b-08f0-4e92-8e4a-b7281f30af3e","Type":"ContainerDied","Data":"82de8f2b633f1b6c295149d077bfd2b308f5acaa429acf411ed16f456536d0ba"} Mar 12 00:11:15 crc kubenswrapper[4948]: I0312 00:11:15.219347 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-rflhd" Mar 12 00:11:15 crc kubenswrapper[4948]: I0312 00:11:15.239977 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-v8h88" Mar 12 00:11:15 crc kubenswrapper[4948]: I0312 00:11:15.290467 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-rflhd" Mar 12 00:11:15 crc kubenswrapper[4948]: I0312 00:11:15.352034 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-62wnf" Mar 12 00:11:15 crc kubenswrapper[4948]: I0312 00:11:15.352373 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-62wnf" Mar 12 00:11:15 crc kubenswrapper[4948]: I0312 00:11:15.402431 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-62wnf" Mar 12 00:11:15 crc kubenswrapper[4948]: I0312 00:11:15.693081 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-hcvm5" Mar 12 00:11:15 crc kubenswrapper[4948]: I0312 00:11:15.693413 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-hcvm5" Mar 12 00:11:15 crc kubenswrapper[4948]: I0312 00:11:15.751575 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-hcvm5" Mar 12 00:11:16 crc kubenswrapper[4948]: I0312 00:11:16.193903 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" event={"ID":"7e483c2b-08f0-4e92-8e4a-b7281f30af3e","Type":"ContainerStarted","Data":"e9ea8517b452a75f7a88c33d420f90b390e28709d728d5fcba49107a76e86013"} Mar 12 00:11:16 crc kubenswrapper[4948]: I0312 00:11:16.250498 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-hcvm5" Mar 12 00:11:16 crc kubenswrapper[4948]: I0312 00:11:16.265572 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-62wnf" Mar 12 00:11:17 crc kubenswrapper[4948]: I0312 00:11:17.644824 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hcvm5"] Mar 12 00:11:17 crc kubenswrapper[4948]: I0312 00:11:17.841999 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-62wnf"] Mar 12 00:11:18 crc kubenswrapper[4948]: I0312 00:11:18.208405 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-hcvm5" podUID="239cfe46-3a3a-4624-9aa2-ea7112f28947" containerName="registry-server" containerID="cri-o://6c529a18c3f1f58c2f5de6e8fd7069a26d1d10de15f10b482d95979542b13de1" gracePeriod=2 Mar 12 00:11:18 crc kubenswrapper[4948]: I0312 00:11:18.780497 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hcvm5" Mar 12 00:11:18 crc kubenswrapper[4948]: I0312 00:11:18.959739 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/239cfe46-3a3a-4624-9aa2-ea7112f28947-utilities\") pod \"239cfe46-3a3a-4624-9aa2-ea7112f28947\" (UID: \"239cfe46-3a3a-4624-9aa2-ea7112f28947\") " Mar 12 00:11:18 crc kubenswrapper[4948]: I0312 00:11:18.959850 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-pk6rv\" (UniqueName: \"kubernetes.io/projected/239cfe46-3a3a-4624-9aa2-ea7112f28947-kube-api-access-pk6rv\") pod \"239cfe46-3a3a-4624-9aa2-ea7112f28947\" (UID: \"239cfe46-3a3a-4624-9aa2-ea7112f28947\") " Mar 12 00:11:18 crc kubenswrapper[4948]: I0312 00:11:18.959913 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/239cfe46-3a3a-4624-9aa2-ea7112f28947-catalog-content\") pod \"239cfe46-3a3a-4624-9aa2-ea7112f28947\" (UID: \"239cfe46-3a3a-4624-9aa2-ea7112f28947\") " Mar 12 00:11:18 crc kubenswrapper[4948]: I0312 00:11:18.961630 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/239cfe46-3a3a-4624-9aa2-ea7112f28947-utilities" (OuterVolumeSpecName: "utilities") pod "239cfe46-3a3a-4624-9aa2-ea7112f28947" (UID: "239cfe46-3a3a-4624-9aa2-ea7112f28947"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 00:11:18 crc kubenswrapper[4948]: I0312 00:11:18.968197 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/239cfe46-3a3a-4624-9aa2-ea7112f28947-kube-api-access-pk6rv" (OuterVolumeSpecName: "kube-api-access-pk6rv") pod "239cfe46-3a3a-4624-9aa2-ea7112f28947" (UID: "239cfe46-3a3a-4624-9aa2-ea7112f28947"). InnerVolumeSpecName "kube-api-access-pk6rv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:11:19 crc kubenswrapper[4948]: I0312 00:11:19.061580 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-pk6rv\" (UniqueName: \"kubernetes.io/projected/239cfe46-3a3a-4624-9aa2-ea7112f28947-kube-api-access-pk6rv\") on node \"crc\" DevicePath \"\"" Mar 12 00:11:19 crc kubenswrapper[4948]: I0312 00:11:19.061623 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/239cfe46-3a3a-4624-9aa2-ea7112f28947-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 00:11:19 crc kubenswrapper[4948]: I0312 00:11:19.064659 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/239cfe46-3a3a-4624-9aa2-ea7112f28947-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "239cfe46-3a3a-4624-9aa2-ea7112f28947" (UID: "239cfe46-3a3a-4624-9aa2-ea7112f28947"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 00:11:19 crc kubenswrapper[4948]: I0312 00:11:19.163538 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/239cfe46-3a3a-4624-9aa2-ea7112f28947-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 00:11:19 crc kubenswrapper[4948]: I0312 00:11:19.220548 4948 generic.go:334] "Generic (PLEG): container finished" podID="239cfe46-3a3a-4624-9aa2-ea7112f28947" containerID="6c529a18c3f1f58c2f5de6e8fd7069a26d1d10de15f10b482d95979542b13de1" exitCode=0 Mar 12 00:11:19 crc kubenswrapper[4948]: I0312 00:11:19.220864 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-62wnf" podUID="e6318b61-ef2d-41da-8c92-ff7523ce613f" containerName="registry-server" containerID="cri-o://663949e58f30470b63459dc123ff8a880fc6ba09cf46402c77d36e17f4cab6fb" gracePeriod=2 Mar 12 00:11:19 crc kubenswrapper[4948]: I0312 00:11:19.221049 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-hcvm5" Mar 12 00:11:19 crc kubenswrapper[4948]: I0312 00:11:19.221556 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hcvm5" event={"ID":"239cfe46-3a3a-4624-9aa2-ea7112f28947","Type":"ContainerDied","Data":"6c529a18c3f1f58c2f5de6e8fd7069a26d1d10de15f10b482d95979542b13de1"} Mar 12 00:11:19 crc kubenswrapper[4948]: I0312 00:11:19.221678 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-hcvm5" event={"ID":"239cfe46-3a3a-4624-9aa2-ea7112f28947","Type":"ContainerDied","Data":"0cf8f970926697cdfac352fdf93f8628843806a504f3e5fcb5654e9560c5011d"} Mar 12 00:11:19 crc kubenswrapper[4948]: I0312 00:11:19.221735 4948 scope.go:117] "RemoveContainer" containerID="6c529a18c3f1f58c2f5de6e8fd7069a26d1d10de15f10b482d95979542b13de1" Mar 12 00:11:19 crc kubenswrapper[4948]: I0312 00:11:19.252667 4948 scope.go:117] "RemoveContainer" containerID="1773b11c62a5767b2972ce020f88a29bd0e48a9ebb0cad24731ba22620298d87" Mar 12 00:11:19 crc kubenswrapper[4948]: I0312 00:11:19.275825 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-hcvm5"] Mar 12 00:11:19 crc kubenswrapper[4948]: I0312 00:11:19.280745 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-hcvm5"] Mar 12 00:11:19 crc kubenswrapper[4948]: I0312 00:11:19.303843 4948 scope.go:117] "RemoveContainer" containerID="9bad20e9704e9652c189e2d25b211eb7431aaf5b1083932c44e8b9e48b7c1dc7" Mar 12 00:11:19 crc kubenswrapper[4948]: I0312 00:11:19.324807 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="239cfe46-3a3a-4624-9aa2-ea7112f28947" path="/var/lib/kubelet/pods/239cfe46-3a3a-4624-9aa2-ea7112f28947/volumes" Mar 12 00:11:19 crc kubenswrapper[4948]: I0312 00:11:19.369544 4948 scope.go:117] "RemoveContainer" containerID="6c529a18c3f1f58c2f5de6e8fd7069a26d1d10de15f10b482d95979542b13de1" Mar 12 00:11:19 crc kubenswrapper[4948]: E0312 00:11:19.370113 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"6c529a18c3f1f58c2f5de6e8fd7069a26d1d10de15f10b482d95979542b13de1\": container with ID starting with 6c529a18c3f1f58c2f5de6e8fd7069a26d1d10de15f10b482d95979542b13de1 not found: ID does not exist" containerID="6c529a18c3f1f58c2f5de6e8fd7069a26d1d10de15f10b482d95979542b13de1" Mar 12 00:11:19 crc kubenswrapper[4948]: I0312 00:11:19.370154 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"6c529a18c3f1f58c2f5de6e8fd7069a26d1d10de15f10b482d95979542b13de1"} err="failed to get container status \"6c529a18c3f1f58c2f5de6e8fd7069a26d1d10de15f10b482d95979542b13de1\": rpc error: code = NotFound desc = could not find container \"6c529a18c3f1f58c2f5de6e8fd7069a26d1d10de15f10b482d95979542b13de1\": container with ID starting with 6c529a18c3f1f58c2f5de6e8fd7069a26d1d10de15f10b482d95979542b13de1 not found: ID does not exist" Mar 12 00:11:19 crc kubenswrapper[4948]: I0312 00:11:19.370187 4948 scope.go:117] "RemoveContainer" containerID="1773b11c62a5767b2972ce020f88a29bd0e48a9ebb0cad24731ba22620298d87" Mar 12 00:11:19 crc kubenswrapper[4948]: E0312 00:11:19.370674 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1773b11c62a5767b2972ce020f88a29bd0e48a9ebb0cad24731ba22620298d87\": container with ID starting with 1773b11c62a5767b2972ce020f88a29bd0e48a9ebb0cad24731ba22620298d87 not found: ID does not exist" containerID="1773b11c62a5767b2972ce020f88a29bd0e48a9ebb0cad24731ba22620298d87" Mar 12 00:11:19 crc kubenswrapper[4948]: I0312 00:11:19.370709 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1773b11c62a5767b2972ce020f88a29bd0e48a9ebb0cad24731ba22620298d87"} err="failed to get container status \"1773b11c62a5767b2972ce020f88a29bd0e48a9ebb0cad24731ba22620298d87\": rpc error: code = NotFound desc = could not find container \"1773b11c62a5767b2972ce020f88a29bd0e48a9ebb0cad24731ba22620298d87\": container with ID starting with 1773b11c62a5767b2972ce020f88a29bd0e48a9ebb0cad24731ba22620298d87 not found: ID does not exist" Mar 12 00:11:19 crc kubenswrapper[4948]: I0312 00:11:19.370743 4948 scope.go:117] "RemoveContainer" containerID="9bad20e9704e9652c189e2d25b211eb7431aaf5b1083932c44e8b9e48b7c1dc7" Mar 12 00:11:19 crc kubenswrapper[4948]: E0312 00:11:19.371083 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"9bad20e9704e9652c189e2d25b211eb7431aaf5b1083932c44e8b9e48b7c1dc7\": container with ID starting with 9bad20e9704e9652c189e2d25b211eb7431aaf5b1083932c44e8b9e48b7c1dc7 not found: ID does not exist" containerID="9bad20e9704e9652c189e2d25b211eb7431aaf5b1083932c44e8b9e48b7c1dc7" Mar 12 00:11:19 crc kubenswrapper[4948]: I0312 00:11:19.371110 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"9bad20e9704e9652c189e2d25b211eb7431aaf5b1083932c44e8b9e48b7c1dc7"} err="failed to get container status \"9bad20e9704e9652c189e2d25b211eb7431aaf5b1083932c44e8b9e48b7c1dc7\": rpc error: code = NotFound desc = could not find container \"9bad20e9704e9652c189e2d25b211eb7431aaf5b1083932c44e8b9e48b7c1dc7\": container with ID starting with 9bad20e9704e9652c189e2d25b211eb7431aaf5b1083932c44e8b9e48b7c1dc7 not found: ID does not exist" Mar 12 00:11:19 crc kubenswrapper[4948]: I0312 00:11:19.779962 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-62wnf" Mar 12 00:11:19 crc kubenswrapper[4948]: I0312 00:11:19.879731 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-5gmsv\" (UniqueName: \"kubernetes.io/projected/e6318b61-ef2d-41da-8c92-ff7523ce613f-kube-api-access-5gmsv\") pod \"e6318b61-ef2d-41da-8c92-ff7523ce613f\" (UID: \"e6318b61-ef2d-41da-8c92-ff7523ce613f\") " Mar 12 00:11:19 crc kubenswrapper[4948]: I0312 00:11:19.880832 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6318b61-ef2d-41da-8c92-ff7523ce613f-utilities\") pod \"e6318b61-ef2d-41da-8c92-ff7523ce613f\" (UID: \"e6318b61-ef2d-41da-8c92-ff7523ce613f\") " Mar 12 00:11:19 crc kubenswrapper[4948]: I0312 00:11:19.881648 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6318b61-ef2d-41da-8c92-ff7523ce613f-catalog-content\") pod \"e6318b61-ef2d-41da-8c92-ff7523ce613f\" (UID: \"e6318b61-ef2d-41da-8c92-ff7523ce613f\") " Mar 12 00:11:19 crc kubenswrapper[4948]: I0312 00:11:19.882126 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e6318b61-ef2d-41da-8c92-ff7523ce613f-utilities" (OuterVolumeSpecName: "utilities") pod "e6318b61-ef2d-41da-8c92-ff7523ce613f" (UID: "e6318b61-ef2d-41da-8c92-ff7523ce613f"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 00:11:19 crc kubenswrapper[4948]: I0312 00:11:19.901499 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e6318b61-ef2d-41da-8c92-ff7523ce613f-kube-api-access-5gmsv" (OuterVolumeSpecName: "kube-api-access-5gmsv") pod "e6318b61-ef2d-41da-8c92-ff7523ce613f" (UID: "e6318b61-ef2d-41da-8c92-ff7523ce613f"). InnerVolumeSpecName "kube-api-access-5gmsv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:11:19 crc kubenswrapper[4948]: I0312 00:11:19.942340 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/e6318b61-ef2d-41da-8c92-ff7523ce613f-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "e6318b61-ef2d-41da-8c92-ff7523ce613f" (UID: "e6318b61-ef2d-41da-8c92-ff7523ce613f"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 00:11:19 crc kubenswrapper[4948]: I0312 00:11:19.983203 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/e6318b61-ef2d-41da-8c92-ff7523ce613f-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 00:11:19 crc kubenswrapper[4948]: I0312 00:11:19.983419 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-5gmsv\" (UniqueName: \"kubernetes.io/projected/e6318b61-ef2d-41da-8c92-ff7523ce613f-kube-api-access-5gmsv\") on node \"crc\" DevicePath \"\"" Mar 12 00:11:19 crc kubenswrapper[4948]: I0312 00:11:19.983491 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/e6318b61-ef2d-41da-8c92-ff7523ce613f-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 00:11:20 crc kubenswrapper[4948]: I0312 00:11:20.234771 4948 generic.go:334] "Generic (PLEG): container finished" podID="e6318b61-ef2d-41da-8c92-ff7523ce613f" containerID="663949e58f30470b63459dc123ff8a880fc6ba09cf46402c77d36e17f4cab6fb" exitCode=0 Mar 12 00:11:20 crc kubenswrapper[4948]: I0312 00:11:20.234836 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-62wnf" event={"ID":"e6318b61-ef2d-41da-8c92-ff7523ce613f","Type":"ContainerDied","Data":"663949e58f30470b63459dc123ff8a880fc6ba09cf46402c77d36e17f4cab6fb"} Mar 12 00:11:20 crc kubenswrapper[4948]: I0312 00:11:20.234880 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-62wnf" Mar 12 00:11:20 crc kubenswrapper[4948]: I0312 00:11:20.234904 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-62wnf" event={"ID":"e6318b61-ef2d-41da-8c92-ff7523ce613f","Type":"ContainerDied","Data":"470ffe54e5d1194c3f772e05a00e8e8d3d874f5fac0e64b8ffad2c49d79ba71f"} Mar 12 00:11:20 crc kubenswrapper[4948]: I0312 00:11:20.234962 4948 scope.go:117] "RemoveContainer" containerID="663949e58f30470b63459dc123ff8a880fc6ba09cf46402c77d36e17f4cab6fb" Mar 12 00:11:20 crc kubenswrapper[4948]: I0312 00:11:20.259977 4948 scope.go:117] "RemoveContainer" containerID="8848e7cb0bd5b2eee15cd54c7a8a9b39bffb36f27a9fc09e35e7884845d59b5b" Mar 12 00:11:20 crc kubenswrapper[4948]: I0312 00:11:20.281643 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-62wnf"] Mar 12 00:11:20 crc kubenswrapper[4948]: I0312 00:11:20.286833 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-62wnf"] Mar 12 00:11:20 crc kubenswrapper[4948]: I0312 00:11:20.305553 4948 scope.go:117] "RemoveContainer" containerID="bfa1f11178206f64a1411b6b67c539f5cd703dddb5b89197ebf8a1b95ff74ebd" Mar 12 00:11:20 crc kubenswrapper[4948]: I0312 00:11:20.327674 4948 scope.go:117] "RemoveContainer" containerID="663949e58f30470b63459dc123ff8a880fc6ba09cf46402c77d36e17f4cab6fb" Mar 12 00:11:20 crc kubenswrapper[4948]: E0312 00:11:20.328556 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"663949e58f30470b63459dc123ff8a880fc6ba09cf46402c77d36e17f4cab6fb\": container with ID starting with 663949e58f30470b63459dc123ff8a880fc6ba09cf46402c77d36e17f4cab6fb not found: ID does not exist" containerID="663949e58f30470b63459dc123ff8a880fc6ba09cf46402c77d36e17f4cab6fb" Mar 12 00:11:20 crc kubenswrapper[4948]: I0312 00:11:20.328638 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"663949e58f30470b63459dc123ff8a880fc6ba09cf46402c77d36e17f4cab6fb"} err="failed to get container status \"663949e58f30470b63459dc123ff8a880fc6ba09cf46402c77d36e17f4cab6fb\": rpc error: code = NotFound desc = could not find container \"663949e58f30470b63459dc123ff8a880fc6ba09cf46402c77d36e17f4cab6fb\": container with ID starting with 663949e58f30470b63459dc123ff8a880fc6ba09cf46402c77d36e17f4cab6fb not found: ID does not exist" Mar 12 00:11:20 crc kubenswrapper[4948]: I0312 00:11:20.328681 4948 scope.go:117] "RemoveContainer" containerID="8848e7cb0bd5b2eee15cd54c7a8a9b39bffb36f27a9fc09e35e7884845d59b5b" Mar 12 00:11:20 crc kubenswrapper[4948]: E0312 00:11:20.329262 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8848e7cb0bd5b2eee15cd54c7a8a9b39bffb36f27a9fc09e35e7884845d59b5b\": container with ID starting with 8848e7cb0bd5b2eee15cd54c7a8a9b39bffb36f27a9fc09e35e7884845d59b5b not found: ID does not exist" containerID="8848e7cb0bd5b2eee15cd54c7a8a9b39bffb36f27a9fc09e35e7884845d59b5b" Mar 12 00:11:20 crc kubenswrapper[4948]: I0312 00:11:20.329356 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8848e7cb0bd5b2eee15cd54c7a8a9b39bffb36f27a9fc09e35e7884845d59b5b"} err="failed to get container status \"8848e7cb0bd5b2eee15cd54c7a8a9b39bffb36f27a9fc09e35e7884845d59b5b\": rpc error: code = NotFound desc = could not find container \"8848e7cb0bd5b2eee15cd54c7a8a9b39bffb36f27a9fc09e35e7884845d59b5b\": container with ID starting with 8848e7cb0bd5b2eee15cd54c7a8a9b39bffb36f27a9fc09e35e7884845d59b5b not found: ID does not exist" Mar 12 00:11:20 crc kubenswrapper[4948]: I0312 00:11:20.329389 4948 scope.go:117] "RemoveContainer" containerID="bfa1f11178206f64a1411b6b67c539f5cd703dddb5b89197ebf8a1b95ff74ebd" Mar 12 00:11:20 crc kubenswrapper[4948]: E0312 00:11:20.329941 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"bfa1f11178206f64a1411b6b67c539f5cd703dddb5b89197ebf8a1b95ff74ebd\": container with ID starting with bfa1f11178206f64a1411b6b67c539f5cd703dddb5b89197ebf8a1b95ff74ebd not found: ID does not exist" containerID="bfa1f11178206f64a1411b6b67c539f5cd703dddb5b89197ebf8a1b95ff74ebd" Mar 12 00:11:20 crc kubenswrapper[4948]: I0312 00:11:20.329998 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"bfa1f11178206f64a1411b6b67c539f5cd703dddb5b89197ebf8a1b95ff74ebd"} err="failed to get container status \"bfa1f11178206f64a1411b6b67c539f5cd703dddb5b89197ebf8a1b95ff74ebd\": rpc error: code = NotFound desc = could not find container \"bfa1f11178206f64a1411b6b67c539f5cd703dddb5b89197ebf8a1b95ff74ebd\": container with ID starting with bfa1f11178206f64a1411b6b67c539f5cd703dddb5b89197ebf8a1b95ff74ebd not found: ID does not exist" Mar 12 00:11:21 crc kubenswrapper[4948]: I0312 00:11:21.317342 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="e6318b61-ef2d-41da-8c92-ff7523ce613f" path="/var/lib/kubelet/pods/e6318b61-ef2d-41da-8c92-ff7523ce613f/volumes" Mar 12 00:11:21 crc kubenswrapper[4948]: I0312 00:11:21.348396 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7f5b4654f4-jnbsn"] Mar 12 00:11:21 crc kubenswrapper[4948]: I0312 00:11:21.348599 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-7f5b4654f4-jnbsn" podUID="25bd67ae-99ad-4f13-a87f-8416e060d3b2" containerName="controller-manager" containerID="cri-o://c5cdf1ba032993b1192bfec19272b4339c9264b00e52452d109a56007da854f8" gracePeriod=30 Mar 12 00:11:21 crc kubenswrapper[4948]: I0312 00:11:21.370530 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-66dcfbd4c4-7wvzp"] Mar 12 00:11:21 crc kubenswrapper[4948]: I0312 00:11:21.370724 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-66dcfbd4c4-7wvzp" podUID="23f23bd1-c3ee-4ec7-9abb-ca6d2b3d0deb" containerName="route-controller-manager" containerID="cri-o://c0099690c0154a146b75cfae69df058dd3e6619ff0ccd594576b65e9783f1aaa" gracePeriod=30 Mar 12 00:11:21 crc kubenswrapper[4948]: I0312 00:11:21.866155 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-66dcfbd4c4-7wvzp" Mar 12 00:11:21 crc kubenswrapper[4948]: I0312 00:11:21.920630 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7f5b4654f4-jnbsn" Mar 12 00:11:22 crc kubenswrapper[4948]: I0312 00:11:22.009439 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-d9r2j\" (UniqueName: \"kubernetes.io/projected/23f23bd1-c3ee-4ec7-9abb-ca6d2b3d0deb-kube-api-access-d9r2j\") pod \"23f23bd1-c3ee-4ec7-9abb-ca6d2b3d0deb\" (UID: \"23f23bd1-c3ee-4ec7-9abb-ca6d2b3d0deb\") " Mar 12 00:11:22 crc kubenswrapper[4948]: I0312 00:11:22.009520 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-b97hz\" (UniqueName: \"kubernetes.io/projected/25bd67ae-99ad-4f13-a87f-8416e060d3b2-kube-api-access-b97hz\") pod \"25bd67ae-99ad-4f13-a87f-8416e060d3b2\" (UID: \"25bd67ae-99ad-4f13-a87f-8416e060d3b2\") " Mar 12 00:11:22 crc kubenswrapper[4948]: I0312 00:11:22.009559 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/25bd67ae-99ad-4f13-a87f-8416e060d3b2-client-ca\") pod \"25bd67ae-99ad-4f13-a87f-8416e060d3b2\" (UID: \"25bd67ae-99ad-4f13-a87f-8416e060d3b2\") " Mar 12 00:11:22 crc kubenswrapper[4948]: I0312 00:11:22.009586 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/23f23bd1-c3ee-4ec7-9abb-ca6d2b3d0deb-client-ca\") pod \"23f23bd1-c3ee-4ec7-9abb-ca6d2b3d0deb\" (UID: \"23f23bd1-c3ee-4ec7-9abb-ca6d2b3d0deb\") " Mar 12 00:11:22 crc kubenswrapper[4948]: I0312 00:11:22.009620 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/25bd67ae-99ad-4f13-a87f-8416e060d3b2-serving-cert\") pod \"25bd67ae-99ad-4f13-a87f-8416e060d3b2\" (UID: \"25bd67ae-99ad-4f13-a87f-8416e060d3b2\") " Mar 12 00:11:22 crc kubenswrapper[4948]: I0312 00:11:22.009660 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25bd67ae-99ad-4f13-a87f-8416e060d3b2-config\") pod \"25bd67ae-99ad-4f13-a87f-8416e060d3b2\" (UID: \"25bd67ae-99ad-4f13-a87f-8416e060d3b2\") " Mar 12 00:11:22 crc kubenswrapper[4948]: I0312 00:11:22.009680 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/23f23bd1-c3ee-4ec7-9abb-ca6d2b3d0deb-serving-cert\") pod \"23f23bd1-c3ee-4ec7-9abb-ca6d2b3d0deb\" (UID: \"23f23bd1-c3ee-4ec7-9abb-ca6d2b3d0deb\") " Mar 12 00:11:22 crc kubenswrapper[4948]: I0312 00:11:22.009703 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/25bd67ae-99ad-4f13-a87f-8416e060d3b2-proxy-ca-bundles\") pod \"25bd67ae-99ad-4f13-a87f-8416e060d3b2\" (UID: \"25bd67ae-99ad-4f13-a87f-8416e060d3b2\") " Mar 12 00:11:22 crc kubenswrapper[4948]: I0312 00:11:22.009723 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23f23bd1-c3ee-4ec7-9abb-ca6d2b3d0deb-config\") pod \"23f23bd1-c3ee-4ec7-9abb-ca6d2b3d0deb\" (UID: \"23f23bd1-c3ee-4ec7-9abb-ca6d2b3d0deb\") " Mar 12 00:11:22 crc kubenswrapper[4948]: I0312 00:11:22.011280 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25bd67ae-99ad-4f13-a87f-8416e060d3b2-client-ca" (OuterVolumeSpecName: "client-ca") pod "25bd67ae-99ad-4f13-a87f-8416e060d3b2" (UID: "25bd67ae-99ad-4f13-a87f-8416e060d3b2"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:11:22 crc kubenswrapper[4948]: I0312 00:11:22.011433 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/23f23bd1-c3ee-4ec7-9abb-ca6d2b3d0deb-client-ca" (OuterVolumeSpecName: "client-ca") pod "23f23bd1-c3ee-4ec7-9abb-ca6d2b3d0deb" (UID: "23f23bd1-c3ee-4ec7-9abb-ca6d2b3d0deb"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:11:22 crc kubenswrapper[4948]: I0312 00:11:22.011521 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/23f23bd1-c3ee-4ec7-9abb-ca6d2b3d0deb-config" (OuterVolumeSpecName: "config") pod "23f23bd1-c3ee-4ec7-9abb-ca6d2b3d0deb" (UID: "23f23bd1-c3ee-4ec7-9abb-ca6d2b3d0deb"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:11:22 crc kubenswrapper[4948]: I0312 00:11:22.011551 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25bd67ae-99ad-4f13-a87f-8416e060d3b2-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "25bd67ae-99ad-4f13-a87f-8416e060d3b2" (UID: "25bd67ae-99ad-4f13-a87f-8416e060d3b2"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:11:22 crc kubenswrapper[4948]: I0312 00:11:22.011633 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/25bd67ae-99ad-4f13-a87f-8416e060d3b2-config" (OuterVolumeSpecName: "config") pod "25bd67ae-99ad-4f13-a87f-8416e060d3b2" (UID: "25bd67ae-99ad-4f13-a87f-8416e060d3b2"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:11:22 crc kubenswrapper[4948]: I0312 00:11:22.017478 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/23f23bd1-c3ee-4ec7-9abb-ca6d2b3d0deb-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "23f23bd1-c3ee-4ec7-9abb-ca6d2b3d0deb" (UID: "23f23bd1-c3ee-4ec7-9abb-ca6d2b3d0deb"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:11:22 crc kubenswrapper[4948]: I0312 00:11:22.017620 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/23f23bd1-c3ee-4ec7-9abb-ca6d2b3d0deb-kube-api-access-d9r2j" (OuterVolumeSpecName: "kube-api-access-d9r2j") pod "23f23bd1-c3ee-4ec7-9abb-ca6d2b3d0deb" (UID: "23f23bd1-c3ee-4ec7-9abb-ca6d2b3d0deb"). InnerVolumeSpecName "kube-api-access-d9r2j". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:11:22 crc kubenswrapper[4948]: I0312 00:11:22.017658 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/25bd67ae-99ad-4f13-a87f-8416e060d3b2-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "25bd67ae-99ad-4f13-a87f-8416e060d3b2" (UID: "25bd67ae-99ad-4f13-a87f-8416e060d3b2"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:11:22 crc kubenswrapper[4948]: I0312 00:11:22.017689 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25bd67ae-99ad-4f13-a87f-8416e060d3b2-kube-api-access-b97hz" (OuterVolumeSpecName: "kube-api-access-b97hz") pod "25bd67ae-99ad-4f13-a87f-8416e060d3b2" (UID: "25bd67ae-99ad-4f13-a87f-8416e060d3b2"). InnerVolumeSpecName "kube-api-access-b97hz". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:11:22 crc kubenswrapper[4948]: I0312 00:11:22.110753 4948 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/23f23bd1-c3ee-4ec7-9abb-ca6d2b3d0deb-client-ca\") on node \"crc\" DevicePath \"\"" Mar 12 00:11:22 crc kubenswrapper[4948]: I0312 00:11:22.110790 4948 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/25bd67ae-99ad-4f13-a87f-8416e060d3b2-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 00:11:22 crc kubenswrapper[4948]: I0312 00:11:22.110800 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/25bd67ae-99ad-4f13-a87f-8416e060d3b2-config\") on node \"crc\" DevicePath \"\"" Mar 12 00:11:22 crc kubenswrapper[4948]: I0312 00:11:22.110808 4948 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/23f23bd1-c3ee-4ec7-9abb-ca6d2b3d0deb-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 00:11:22 crc kubenswrapper[4948]: I0312 00:11:22.110817 4948 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/25bd67ae-99ad-4f13-a87f-8416e060d3b2-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 12 00:11:22 crc kubenswrapper[4948]: I0312 00:11:22.110826 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/23f23bd1-c3ee-4ec7-9abb-ca6d2b3d0deb-config\") on node \"crc\" DevicePath \"\"" Mar 12 00:11:22 crc kubenswrapper[4948]: I0312 00:11:22.110834 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-d9r2j\" (UniqueName: \"kubernetes.io/projected/23f23bd1-c3ee-4ec7-9abb-ca6d2b3d0deb-kube-api-access-d9r2j\") on node \"crc\" DevicePath \"\"" Mar 12 00:11:22 crc kubenswrapper[4948]: I0312 00:11:22.110844 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-b97hz\" (UniqueName: \"kubernetes.io/projected/25bd67ae-99ad-4f13-a87f-8416e060d3b2-kube-api-access-b97hz\") on node \"crc\" DevicePath \"\"" Mar 12 00:11:22 crc kubenswrapper[4948]: I0312 00:11:22.110853 4948 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/25bd67ae-99ad-4f13-a87f-8416e060d3b2-client-ca\") on node \"crc\" DevicePath \"\"" Mar 12 00:11:22 crc kubenswrapper[4948]: I0312 00:11:22.254224 4948 generic.go:334] "Generic (PLEG): container finished" podID="25bd67ae-99ad-4f13-a87f-8416e060d3b2" containerID="c5cdf1ba032993b1192bfec19272b4339c9264b00e52452d109a56007da854f8" exitCode=0 Mar 12 00:11:22 crc kubenswrapper[4948]: I0312 00:11:22.254272 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7f5b4654f4-jnbsn" event={"ID":"25bd67ae-99ad-4f13-a87f-8416e060d3b2","Type":"ContainerDied","Data":"c5cdf1ba032993b1192bfec19272b4339c9264b00e52452d109a56007da854f8"} Mar 12 00:11:22 crc kubenswrapper[4948]: I0312 00:11:22.254335 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-7f5b4654f4-jnbsn" event={"ID":"25bd67ae-99ad-4f13-a87f-8416e060d3b2","Type":"ContainerDied","Data":"d2ea3333332ca69448406b292cd34c2c9cbd23b39cfa77b04ef926844caabf84"} Mar 12 00:11:22 crc kubenswrapper[4948]: I0312 00:11:22.254354 4948 scope.go:117] "RemoveContainer" containerID="c5cdf1ba032993b1192bfec19272b4339c9264b00e52452d109a56007da854f8" Mar 12 00:11:22 crc kubenswrapper[4948]: I0312 00:11:22.254376 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-7f5b4654f4-jnbsn" Mar 12 00:11:22 crc kubenswrapper[4948]: I0312 00:11:22.256741 4948 generic.go:334] "Generic (PLEG): container finished" podID="23f23bd1-c3ee-4ec7-9abb-ca6d2b3d0deb" containerID="c0099690c0154a146b75cfae69df058dd3e6619ff0ccd594576b65e9783f1aaa" exitCode=0 Mar 12 00:11:22 crc kubenswrapper[4948]: I0312 00:11:22.256804 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-66dcfbd4c4-7wvzp" event={"ID":"23f23bd1-c3ee-4ec7-9abb-ca6d2b3d0deb","Type":"ContainerDied","Data":"c0099690c0154a146b75cfae69df058dd3e6619ff0ccd594576b65e9783f1aaa"} Mar 12 00:11:22 crc kubenswrapper[4948]: I0312 00:11:22.256846 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-66dcfbd4c4-7wvzp" event={"ID":"23f23bd1-c3ee-4ec7-9abb-ca6d2b3d0deb","Type":"ContainerDied","Data":"5c8e0932de21699e210003c636cf194223dcb9522b78b078b934452229261336"} Mar 12 00:11:22 crc kubenswrapper[4948]: I0312 00:11:22.256811 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-66dcfbd4c4-7wvzp" Mar 12 00:11:22 crc kubenswrapper[4948]: I0312 00:11:22.278362 4948 scope.go:117] "RemoveContainer" containerID="c5cdf1ba032993b1192bfec19272b4339c9264b00e52452d109a56007da854f8" Mar 12 00:11:22 crc kubenswrapper[4948]: E0312 00:11:22.278854 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c5cdf1ba032993b1192bfec19272b4339c9264b00e52452d109a56007da854f8\": container with ID starting with c5cdf1ba032993b1192bfec19272b4339c9264b00e52452d109a56007da854f8 not found: ID does not exist" containerID="c5cdf1ba032993b1192bfec19272b4339c9264b00e52452d109a56007da854f8" Mar 12 00:11:22 crc kubenswrapper[4948]: I0312 00:11:22.278904 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c5cdf1ba032993b1192bfec19272b4339c9264b00e52452d109a56007da854f8"} err="failed to get container status \"c5cdf1ba032993b1192bfec19272b4339c9264b00e52452d109a56007da854f8\": rpc error: code = NotFound desc = could not find container \"c5cdf1ba032993b1192bfec19272b4339c9264b00e52452d109a56007da854f8\": container with ID starting with c5cdf1ba032993b1192bfec19272b4339c9264b00e52452d109a56007da854f8 not found: ID does not exist" Mar 12 00:11:22 crc kubenswrapper[4948]: I0312 00:11:22.278938 4948 scope.go:117] "RemoveContainer" containerID="c0099690c0154a146b75cfae69df058dd3e6619ff0ccd594576b65e9783f1aaa" Mar 12 00:11:22 crc kubenswrapper[4948]: I0312 00:11:22.294562 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-7f5b4654f4-jnbsn"] Mar 12 00:11:22 crc kubenswrapper[4948]: I0312 00:11:22.304001 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-7f5b4654f4-jnbsn"] Mar 12 00:11:22 crc kubenswrapper[4948]: I0312 00:11:22.306658 4948 scope.go:117] "RemoveContainer" containerID="c0099690c0154a146b75cfae69df058dd3e6619ff0ccd594576b65e9783f1aaa" Mar 12 00:11:22 crc kubenswrapper[4948]: E0312 00:11:22.307426 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c0099690c0154a146b75cfae69df058dd3e6619ff0ccd594576b65e9783f1aaa\": container with ID starting with c0099690c0154a146b75cfae69df058dd3e6619ff0ccd594576b65e9783f1aaa not found: ID does not exist" containerID="c0099690c0154a146b75cfae69df058dd3e6619ff0ccd594576b65e9783f1aaa" Mar 12 00:11:22 crc kubenswrapper[4948]: I0312 00:11:22.307464 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c0099690c0154a146b75cfae69df058dd3e6619ff0ccd594576b65e9783f1aaa"} err="failed to get container status \"c0099690c0154a146b75cfae69df058dd3e6619ff0ccd594576b65e9783f1aaa\": rpc error: code = NotFound desc = could not find container \"c0099690c0154a146b75cfae69df058dd3e6619ff0ccd594576b65e9783f1aaa\": container with ID starting with c0099690c0154a146b75cfae69df058dd3e6619ff0ccd594576b65e9783f1aaa not found: ID does not exist" Mar 12 00:11:22 crc kubenswrapper[4948]: I0312 00:11:22.314377 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-66dcfbd4c4-7wvzp"] Mar 12 00:11:22 crc kubenswrapper[4948]: I0312 00:11:22.320385 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-66dcfbd4c4-7wvzp"] Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.281507 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-76fbfc78f5-ll5hs"] Mar 12 00:11:23 crc kubenswrapper[4948]: E0312 00:11:23.281986 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6318b61-ef2d-41da-8c92-ff7523ce613f" containerName="extract-utilities" Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.282032 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6318b61-ef2d-41da-8c92-ff7523ce613f" containerName="extract-utilities" Mar 12 00:11:23 crc kubenswrapper[4948]: E0312 00:11:23.282054 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="529630b8-cc23-44b7-bafd-6ae1313877d1" containerName="extract-content" Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.282070 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="529630b8-cc23-44b7-bafd-6ae1313877d1" containerName="extract-content" Mar 12 00:11:23 crc kubenswrapper[4948]: E0312 00:11:23.282099 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6318b61-ef2d-41da-8c92-ff7523ce613f" containerName="extract-content" Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.282118 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6318b61-ef2d-41da-8c92-ff7523ce613f" containerName="extract-content" Mar 12 00:11:23 crc kubenswrapper[4948]: E0312 00:11:23.282149 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="23f23bd1-c3ee-4ec7-9abb-ca6d2b3d0deb" containerName="route-controller-manager" Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.282166 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="23f23bd1-c3ee-4ec7-9abb-ca6d2b3d0deb" containerName="route-controller-manager" Mar 12 00:11:23 crc kubenswrapper[4948]: E0312 00:11:23.282187 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44212e8e-c675-4643-a752-997be5f8e980" containerName="extract-utilities" Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.282202 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="44212e8e-c675-4643-a752-997be5f8e980" containerName="extract-utilities" Mar 12 00:11:23 crc kubenswrapper[4948]: E0312 00:11:23.282266 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="529630b8-cc23-44b7-bafd-6ae1313877d1" containerName="registry-server" Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.282284 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="529630b8-cc23-44b7-bafd-6ae1313877d1" containerName="registry-server" Mar 12 00:11:23 crc kubenswrapper[4948]: E0312 00:11:23.282342 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44212e8e-c675-4643-a752-997be5f8e980" containerName="registry-server" Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.282360 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="44212e8e-c675-4643-a752-997be5f8e980" containerName="registry-server" Mar 12 00:11:23 crc kubenswrapper[4948]: E0312 00:11:23.282384 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e6318b61-ef2d-41da-8c92-ff7523ce613f" containerName="registry-server" Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.282402 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="e6318b61-ef2d-41da-8c92-ff7523ce613f" containerName="registry-server" Mar 12 00:11:23 crc kubenswrapper[4948]: E0312 00:11:23.282427 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="529630b8-cc23-44b7-bafd-6ae1313877d1" containerName="extract-utilities" Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.282443 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="529630b8-cc23-44b7-bafd-6ae1313877d1" containerName="extract-utilities" Mar 12 00:11:23 crc kubenswrapper[4948]: E0312 00:11:23.282469 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="25bd67ae-99ad-4f13-a87f-8416e060d3b2" containerName="controller-manager" Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.282486 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="25bd67ae-99ad-4f13-a87f-8416e060d3b2" containerName="controller-manager" Mar 12 00:11:23 crc kubenswrapper[4948]: E0312 00:11:23.282514 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="239cfe46-3a3a-4624-9aa2-ea7112f28947" containerName="extract-content" Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.282532 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="239cfe46-3a3a-4624-9aa2-ea7112f28947" containerName="extract-content" Mar 12 00:11:23 crc kubenswrapper[4948]: E0312 00:11:23.282561 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="239cfe46-3a3a-4624-9aa2-ea7112f28947" containerName="registry-server" Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.282578 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="239cfe46-3a3a-4624-9aa2-ea7112f28947" containerName="registry-server" Mar 12 00:11:23 crc kubenswrapper[4948]: E0312 00:11:23.282602 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="44212e8e-c675-4643-a752-997be5f8e980" containerName="extract-content" Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.282618 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="44212e8e-c675-4643-a752-997be5f8e980" containerName="extract-content" Mar 12 00:11:23 crc kubenswrapper[4948]: E0312 00:11:23.282644 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="239cfe46-3a3a-4624-9aa2-ea7112f28947" containerName="extract-utilities" Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.282661 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="239cfe46-3a3a-4624-9aa2-ea7112f28947" containerName="extract-utilities" Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.282915 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="e6318b61-ef2d-41da-8c92-ff7523ce613f" containerName="registry-server" Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.282953 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="23f23bd1-c3ee-4ec7-9abb-ca6d2b3d0deb" containerName="route-controller-manager" Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.282976 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="25bd67ae-99ad-4f13-a87f-8416e060d3b2" containerName="controller-manager" Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.283006 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="529630b8-cc23-44b7-bafd-6ae1313877d1" containerName="registry-server" Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.283027 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="239cfe46-3a3a-4624-9aa2-ea7112f28947" containerName="registry-server" Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.283048 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="44212e8e-c675-4643-a752-997be5f8e980" containerName="registry-server" Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.283908 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-76fbfc78f5-ll5hs" Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.287555 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5d6f859c69-9lr2f"] Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.288785 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5d6f859c69-9lr2f" Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.291121 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.293773 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.294097 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.294403 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.294621 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.297690 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.297839 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.298155 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.298428 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.298838 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.301095 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.301633 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.307888 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.310021 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-76fbfc78f5-ll5hs"] Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.341124 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="23f23bd1-c3ee-4ec7-9abb-ca6d2b3d0deb" path="/var/lib/kubelet/pods/23f23bd1-c3ee-4ec7-9abb-ca6d2b3d0deb/volumes" Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.342214 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="25bd67ae-99ad-4f13-a87f-8416e060d3b2" path="/var/lib/kubelet/pods/25bd67ae-99ad-4f13-a87f-8416e060d3b2/volumes" Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.343257 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5d6f859c69-9lr2f"] Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.431813 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/122cfac5-b68f-4ac2-9bd3-708c92e07c08-serving-cert\") pod \"route-controller-manager-5d6f859c69-9lr2f\" (UID: \"122cfac5-b68f-4ac2-9bd3-708c92e07c08\") " pod="openshift-route-controller-manager/route-controller-manager-5d6f859c69-9lr2f" Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.432073 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/caf01641-a311-4aac-af83-1edabc201169-serving-cert\") pod \"controller-manager-76fbfc78f5-ll5hs\" (UID: \"caf01641-a311-4aac-af83-1edabc201169\") " pod="openshift-controller-manager/controller-manager-76fbfc78f5-ll5hs" Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.432100 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/122cfac5-b68f-4ac2-9bd3-708c92e07c08-client-ca\") pod \"route-controller-manager-5d6f859c69-9lr2f\" (UID: \"122cfac5-b68f-4ac2-9bd3-708c92e07c08\") " pod="openshift-route-controller-manager/route-controller-manager-5d6f859c69-9lr2f" Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.432118 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-wh6pm\" (UniqueName: \"kubernetes.io/projected/caf01641-a311-4aac-af83-1edabc201169-kube-api-access-wh6pm\") pod \"controller-manager-76fbfc78f5-ll5hs\" (UID: \"caf01641-a311-4aac-af83-1edabc201169\") " pod="openshift-controller-manager/controller-manager-76fbfc78f5-ll5hs" Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.432145 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fmzvt\" (UniqueName: \"kubernetes.io/projected/122cfac5-b68f-4ac2-9bd3-708c92e07c08-kube-api-access-fmzvt\") pod \"route-controller-manager-5d6f859c69-9lr2f\" (UID: \"122cfac5-b68f-4ac2-9bd3-708c92e07c08\") " pod="openshift-route-controller-manager/route-controller-manager-5d6f859c69-9lr2f" Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.432166 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/caf01641-a311-4aac-af83-1edabc201169-proxy-ca-bundles\") pod \"controller-manager-76fbfc78f5-ll5hs\" (UID: \"caf01641-a311-4aac-af83-1edabc201169\") " pod="openshift-controller-manager/controller-manager-76fbfc78f5-ll5hs" Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.432187 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/caf01641-a311-4aac-af83-1edabc201169-client-ca\") pod \"controller-manager-76fbfc78f5-ll5hs\" (UID: \"caf01641-a311-4aac-af83-1edabc201169\") " pod="openshift-controller-manager/controller-manager-76fbfc78f5-ll5hs" Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.432208 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/122cfac5-b68f-4ac2-9bd3-708c92e07c08-config\") pod \"route-controller-manager-5d6f859c69-9lr2f\" (UID: \"122cfac5-b68f-4ac2-9bd3-708c92e07c08\") " pod="openshift-route-controller-manager/route-controller-manager-5d6f859c69-9lr2f" Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.432231 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/caf01641-a311-4aac-af83-1edabc201169-config\") pod \"controller-manager-76fbfc78f5-ll5hs\" (UID: \"caf01641-a311-4aac-af83-1edabc201169\") " pod="openshift-controller-manager/controller-manager-76fbfc78f5-ll5hs" Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.533962 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/caf01641-a311-4aac-af83-1edabc201169-client-ca\") pod \"controller-manager-76fbfc78f5-ll5hs\" (UID: \"caf01641-a311-4aac-af83-1edabc201169\") " pod="openshift-controller-manager/controller-manager-76fbfc78f5-ll5hs" Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.534087 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/122cfac5-b68f-4ac2-9bd3-708c92e07c08-config\") pod \"route-controller-manager-5d6f859c69-9lr2f\" (UID: \"122cfac5-b68f-4ac2-9bd3-708c92e07c08\") " pod="openshift-route-controller-manager/route-controller-manager-5d6f859c69-9lr2f" Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.534138 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/caf01641-a311-4aac-af83-1edabc201169-config\") pod \"controller-manager-76fbfc78f5-ll5hs\" (UID: \"caf01641-a311-4aac-af83-1edabc201169\") " pod="openshift-controller-manager/controller-manager-76fbfc78f5-ll5hs" Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.534202 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/122cfac5-b68f-4ac2-9bd3-708c92e07c08-serving-cert\") pod \"route-controller-manager-5d6f859c69-9lr2f\" (UID: \"122cfac5-b68f-4ac2-9bd3-708c92e07c08\") " pod="openshift-route-controller-manager/route-controller-manager-5d6f859c69-9lr2f" Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.534233 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/caf01641-a311-4aac-af83-1edabc201169-serving-cert\") pod \"controller-manager-76fbfc78f5-ll5hs\" (UID: \"caf01641-a311-4aac-af83-1edabc201169\") " pod="openshift-controller-manager/controller-manager-76fbfc78f5-ll5hs" Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.534282 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/122cfac5-b68f-4ac2-9bd3-708c92e07c08-client-ca\") pod \"route-controller-manager-5d6f859c69-9lr2f\" (UID: \"122cfac5-b68f-4ac2-9bd3-708c92e07c08\") " pod="openshift-route-controller-manager/route-controller-manager-5d6f859c69-9lr2f" Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.534356 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-wh6pm\" (UniqueName: \"kubernetes.io/projected/caf01641-a311-4aac-af83-1edabc201169-kube-api-access-wh6pm\") pod \"controller-manager-76fbfc78f5-ll5hs\" (UID: \"caf01641-a311-4aac-af83-1edabc201169\") " pod="openshift-controller-manager/controller-manager-76fbfc78f5-ll5hs" Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.534402 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fmzvt\" (UniqueName: \"kubernetes.io/projected/122cfac5-b68f-4ac2-9bd3-708c92e07c08-kube-api-access-fmzvt\") pod \"route-controller-manager-5d6f859c69-9lr2f\" (UID: \"122cfac5-b68f-4ac2-9bd3-708c92e07c08\") " pod="openshift-route-controller-manager/route-controller-manager-5d6f859c69-9lr2f" Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.534437 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/caf01641-a311-4aac-af83-1edabc201169-proxy-ca-bundles\") pod \"controller-manager-76fbfc78f5-ll5hs\" (UID: \"caf01641-a311-4aac-af83-1edabc201169\") " pod="openshift-controller-manager/controller-manager-76fbfc78f5-ll5hs" Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.536202 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/caf01641-a311-4aac-af83-1edabc201169-client-ca\") pod \"controller-manager-76fbfc78f5-ll5hs\" (UID: \"caf01641-a311-4aac-af83-1edabc201169\") " pod="openshift-controller-manager/controller-manager-76fbfc78f5-ll5hs" Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.537249 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/122cfac5-b68f-4ac2-9bd3-708c92e07c08-client-ca\") pod \"route-controller-manager-5d6f859c69-9lr2f\" (UID: \"122cfac5-b68f-4ac2-9bd3-708c92e07c08\") " pod="openshift-route-controller-manager/route-controller-manager-5d6f859c69-9lr2f" Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.537512 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/caf01641-a311-4aac-af83-1edabc201169-config\") pod \"controller-manager-76fbfc78f5-ll5hs\" (UID: \"caf01641-a311-4aac-af83-1edabc201169\") " pod="openshift-controller-manager/controller-manager-76fbfc78f5-ll5hs" Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.538243 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/caf01641-a311-4aac-af83-1edabc201169-proxy-ca-bundles\") pod \"controller-manager-76fbfc78f5-ll5hs\" (UID: \"caf01641-a311-4aac-af83-1edabc201169\") " pod="openshift-controller-manager/controller-manager-76fbfc78f5-ll5hs" Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.539184 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/122cfac5-b68f-4ac2-9bd3-708c92e07c08-config\") pod \"route-controller-manager-5d6f859c69-9lr2f\" (UID: \"122cfac5-b68f-4ac2-9bd3-708c92e07c08\") " pod="openshift-route-controller-manager/route-controller-manager-5d6f859c69-9lr2f" Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.544863 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/caf01641-a311-4aac-af83-1edabc201169-serving-cert\") pod \"controller-manager-76fbfc78f5-ll5hs\" (UID: \"caf01641-a311-4aac-af83-1edabc201169\") " pod="openshift-controller-manager/controller-manager-76fbfc78f5-ll5hs" Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.556373 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/122cfac5-b68f-4ac2-9bd3-708c92e07c08-serving-cert\") pod \"route-controller-manager-5d6f859c69-9lr2f\" (UID: \"122cfac5-b68f-4ac2-9bd3-708c92e07c08\") " pod="openshift-route-controller-manager/route-controller-manager-5d6f859c69-9lr2f" Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.566023 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-wh6pm\" (UniqueName: \"kubernetes.io/projected/caf01641-a311-4aac-af83-1edabc201169-kube-api-access-wh6pm\") pod \"controller-manager-76fbfc78f5-ll5hs\" (UID: \"caf01641-a311-4aac-af83-1edabc201169\") " pod="openshift-controller-manager/controller-manager-76fbfc78f5-ll5hs" Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.569577 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fmzvt\" (UniqueName: \"kubernetes.io/projected/122cfac5-b68f-4ac2-9bd3-708c92e07c08-kube-api-access-fmzvt\") pod \"route-controller-manager-5d6f859c69-9lr2f\" (UID: \"122cfac5-b68f-4ac2-9bd3-708c92e07c08\") " pod="openshift-route-controller-manager/route-controller-manager-5d6f859c69-9lr2f" Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.638136 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-76fbfc78f5-ll5hs" Mar 12 00:11:23 crc kubenswrapper[4948]: I0312 00:11:23.652023 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5d6f859c69-9lr2f" Mar 12 00:11:24 crc kubenswrapper[4948]: I0312 00:11:23.999799 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-76fbfc78f5-ll5hs"] Mar 12 00:11:24 crc kubenswrapper[4948]: I0312 00:11:24.146039 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5d6f859c69-9lr2f"] Mar 12 00:11:24 crc kubenswrapper[4948]: I0312 00:11:24.275290 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5d6f859c69-9lr2f" event={"ID":"122cfac5-b68f-4ac2-9bd3-708c92e07c08","Type":"ContainerStarted","Data":"d70fdfbf6ea660cbf04d24d833818679a0c91de9f91e6f689191d9984f783f77"} Mar 12 00:11:24 crc kubenswrapper[4948]: I0312 00:11:24.276910 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-76fbfc78f5-ll5hs" event={"ID":"caf01641-a311-4aac-af83-1edabc201169","Type":"ContainerStarted","Data":"1a7cdabe724b78d08c5ce27a48568c313fa9af595a81c05ec244574a1874f459"} Mar 12 00:11:24 crc kubenswrapper[4948]: I0312 00:11:24.276941 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-76fbfc78f5-ll5hs" event={"ID":"caf01641-a311-4aac-af83-1edabc201169","Type":"ContainerStarted","Data":"daa23c1151891223f66ab40d09e29fa883ebdd8edfbe1155ac83632d20cff03e"} Mar 12 00:11:24 crc kubenswrapper[4948]: I0312 00:11:24.277995 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-76fbfc78f5-ll5hs" Mar 12 00:11:24 crc kubenswrapper[4948]: I0312 00:11:24.282594 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-76fbfc78f5-ll5hs" Mar 12 00:11:24 crc kubenswrapper[4948]: I0312 00:11:24.307915 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-76fbfc78f5-ll5hs" podStartSLOduration=3.307895603 podStartE2EDuration="3.307895603s" podCreationTimestamp="2026-03-12 00:11:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:11:24.304492383 +0000 UTC m=+283.760096141" watchObservedRunningTime="2026-03-12 00:11:24.307895603 +0000 UTC m=+283.763499351" Mar 12 00:11:25 crc kubenswrapper[4948]: I0312 00:11:25.286850 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5d6f859c69-9lr2f" event={"ID":"122cfac5-b68f-4ac2-9bd3-708c92e07c08","Type":"ContainerStarted","Data":"739f4822ffb6114ecbaef0482544cc5ee52e0f0f9d9617e822df4c63b0954128"} Mar 12 00:11:25 crc kubenswrapper[4948]: I0312 00:11:25.316580 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-5d6f859c69-9lr2f" podStartSLOduration=4.316555936 podStartE2EDuration="4.316555936s" podCreationTimestamp="2026-03-12 00:11:21 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:11:25.31207421 +0000 UTC m=+284.767677978" watchObservedRunningTime="2026-03-12 00:11:25.316555936 +0000 UTC m=+284.772159704" Mar 12 00:11:26 crc kubenswrapper[4948]: I0312 00:11:26.291345 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-5d6f859c69-9lr2f" Mar 12 00:11:26 crc kubenswrapper[4948]: I0312 00:11:26.296716 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-5d6f859c69-9lr2f" Mar 12 00:11:26 crc kubenswrapper[4948]: I0312 00:11:26.613186 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-authentication/oauth-openshift-558db77b4-qsf8x"] Mar 12 00:11:41 crc kubenswrapper[4948]: I0312 00:11:41.332121 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-76fbfc78f5-ll5hs"] Mar 12 00:11:41 crc kubenswrapper[4948]: I0312 00:11:41.333957 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-controller-manager/controller-manager-76fbfc78f5-ll5hs" podUID="caf01641-a311-4aac-af83-1edabc201169" containerName="controller-manager" containerID="cri-o://1a7cdabe724b78d08c5ce27a48568c313fa9af595a81c05ec244574a1874f459" gracePeriod=30 Mar 12 00:11:41 crc kubenswrapper[4948]: I0312 00:11:41.428699 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5d6f859c69-9lr2f"] Mar 12 00:11:41 crc kubenswrapper[4948]: I0312 00:11:41.429263 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-route-controller-manager/route-controller-manager-5d6f859c69-9lr2f" podUID="122cfac5-b68f-4ac2-9bd3-708c92e07c08" containerName="route-controller-manager" containerID="cri-o://739f4822ffb6114ecbaef0482544cc5ee52e0f0f9d9617e822df4c63b0954128" gracePeriod=30 Mar 12 00:11:41 crc kubenswrapper[4948]: I0312 00:11:41.938457 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5d6f859c69-9lr2f" Mar 12 00:11:41 crc kubenswrapper[4948]: I0312 00:11:41.944842 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-76fbfc78f5-ll5hs" Mar 12 00:11:41 crc kubenswrapper[4948]: I0312 00:11:41.944946 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/122cfac5-b68f-4ac2-9bd3-708c92e07c08-serving-cert\") pod \"122cfac5-b68f-4ac2-9bd3-708c92e07c08\" (UID: \"122cfac5-b68f-4ac2-9bd3-708c92e07c08\") " Mar 12 00:11:41 crc kubenswrapper[4948]: I0312 00:11:41.945024 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/122cfac5-b68f-4ac2-9bd3-708c92e07c08-client-ca\") pod \"122cfac5-b68f-4ac2-9bd3-708c92e07c08\" (UID: \"122cfac5-b68f-4ac2-9bd3-708c92e07c08\") " Mar 12 00:11:41 crc kubenswrapper[4948]: I0312 00:11:41.945054 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/122cfac5-b68f-4ac2-9bd3-708c92e07c08-config\") pod \"122cfac5-b68f-4ac2-9bd3-708c92e07c08\" (UID: \"122cfac5-b68f-4ac2-9bd3-708c92e07c08\") " Mar 12 00:11:41 crc kubenswrapper[4948]: I0312 00:11:41.945109 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-fmzvt\" (UniqueName: \"kubernetes.io/projected/122cfac5-b68f-4ac2-9bd3-708c92e07c08-kube-api-access-fmzvt\") pod \"122cfac5-b68f-4ac2-9bd3-708c92e07c08\" (UID: \"122cfac5-b68f-4ac2-9bd3-708c92e07c08\") " Mar 12 00:11:41 crc kubenswrapper[4948]: I0312 00:11:41.946739 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/122cfac5-b68f-4ac2-9bd3-708c92e07c08-config" (OuterVolumeSpecName: "config") pod "122cfac5-b68f-4ac2-9bd3-708c92e07c08" (UID: "122cfac5-b68f-4ac2-9bd3-708c92e07c08"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:11:41 crc kubenswrapper[4948]: I0312 00:11:41.946758 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/122cfac5-b68f-4ac2-9bd3-708c92e07c08-client-ca" (OuterVolumeSpecName: "client-ca") pod "122cfac5-b68f-4ac2-9bd3-708c92e07c08" (UID: "122cfac5-b68f-4ac2-9bd3-708c92e07c08"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:11:41 crc kubenswrapper[4948]: I0312 00:11:41.951524 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/122cfac5-b68f-4ac2-9bd3-708c92e07c08-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "122cfac5-b68f-4ac2-9bd3-708c92e07c08" (UID: "122cfac5-b68f-4ac2-9bd3-708c92e07c08"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:11:41 crc kubenswrapper[4948]: I0312 00:11:41.953005 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/122cfac5-b68f-4ac2-9bd3-708c92e07c08-kube-api-access-fmzvt" (OuterVolumeSpecName: "kube-api-access-fmzvt") pod "122cfac5-b68f-4ac2-9bd3-708c92e07c08" (UID: "122cfac5-b68f-4ac2-9bd3-708c92e07c08"). InnerVolumeSpecName "kube-api-access-fmzvt". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:11:42 crc kubenswrapper[4948]: I0312 00:11:42.046263 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-fmzvt\" (UniqueName: \"kubernetes.io/projected/122cfac5-b68f-4ac2-9bd3-708c92e07c08-kube-api-access-fmzvt\") on node \"crc\" DevicePath \"\"" Mar 12 00:11:42 crc kubenswrapper[4948]: I0312 00:11:42.046345 4948 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/122cfac5-b68f-4ac2-9bd3-708c92e07c08-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 00:11:42 crc kubenswrapper[4948]: I0312 00:11:42.046369 4948 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/122cfac5-b68f-4ac2-9bd3-708c92e07c08-client-ca\") on node \"crc\" DevicePath \"\"" Mar 12 00:11:42 crc kubenswrapper[4948]: I0312 00:11:42.046391 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/122cfac5-b68f-4ac2-9bd3-708c92e07c08-config\") on node \"crc\" DevicePath \"\"" Mar 12 00:11:42 crc kubenswrapper[4948]: I0312 00:11:42.147563 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/caf01641-a311-4aac-af83-1edabc201169-client-ca\") pod \"caf01641-a311-4aac-af83-1edabc201169\" (UID: \"caf01641-a311-4aac-af83-1edabc201169\") " Mar 12 00:11:42 crc kubenswrapper[4948]: I0312 00:11:42.147915 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/caf01641-a311-4aac-af83-1edabc201169-config\") pod \"caf01641-a311-4aac-af83-1edabc201169\" (UID: \"caf01641-a311-4aac-af83-1edabc201169\") " Mar 12 00:11:42 crc kubenswrapper[4948]: I0312 00:11:42.148112 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/caf01641-a311-4aac-af83-1edabc201169-proxy-ca-bundles\") pod \"caf01641-a311-4aac-af83-1edabc201169\" (UID: \"caf01641-a311-4aac-af83-1edabc201169\") " Mar 12 00:11:42 crc kubenswrapper[4948]: I0312 00:11:42.148263 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/caf01641-a311-4aac-af83-1edabc201169-client-ca" (OuterVolumeSpecName: "client-ca") pod "caf01641-a311-4aac-af83-1edabc201169" (UID: "caf01641-a311-4aac-af83-1edabc201169"). InnerVolumeSpecName "client-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:11:42 crc kubenswrapper[4948]: I0312 00:11:42.148797 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/caf01641-a311-4aac-af83-1edabc201169-config" (OuterVolumeSpecName: "config") pod "caf01641-a311-4aac-af83-1edabc201169" (UID: "caf01641-a311-4aac-af83-1edabc201169"). InnerVolumeSpecName "config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:11:42 crc kubenswrapper[4948]: I0312 00:11:42.148289 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/caf01641-a311-4aac-af83-1edabc201169-serving-cert\") pod \"caf01641-a311-4aac-af83-1edabc201169\" (UID: \"caf01641-a311-4aac-af83-1edabc201169\") " Mar 12 00:11:42 crc kubenswrapper[4948]: I0312 00:11:42.149181 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-wh6pm\" (UniqueName: \"kubernetes.io/projected/caf01641-a311-4aac-af83-1edabc201169-kube-api-access-wh6pm\") pod \"caf01641-a311-4aac-af83-1edabc201169\" (UID: \"caf01641-a311-4aac-af83-1edabc201169\") " Mar 12 00:11:42 crc kubenswrapper[4948]: I0312 00:11:42.149228 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/caf01641-a311-4aac-af83-1edabc201169-proxy-ca-bundles" (OuterVolumeSpecName: "proxy-ca-bundles") pod "caf01641-a311-4aac-af83-1edabc201169" (UID: "caf01641-a311-4aac-af83-1edabc201169"). InnerVolumeSpecName "proxy-ca-bundles". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:11:42 crc kubenswrapper[4948]: I0312 00:11:42.149769 4948 reconciler_common.go:293] "Volume detached for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/caf01641-a311-4aac-af83-1edabc201169-client-ca\") on node \"crc\" DevicePath \"\"" Mar 12 00:11:42 crc kubenswrapper[4948]: I0312 00:11:42.149883 4948 reconciler_common.go:293] "Volume detached for volume \"config\" (UniqueName: \"kubernetes.io/configmap/caf01641-a311-4aac-af83-1edabc201169-config\") on node \"crc\" DevicePath \"\"" Mar 12 00:11:42 crc kubenswrapper[4948]: I0312 00:11:42.149976 4948 reconciler_common.go:293] "Volume detached for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/caf01641-a311-4aac-af83-1edabc201169-proxy-ca-bundles\") on node \"crc\" DevicePath \"\"" Mar 12 00:11:42 crc kubenswrapper[4948]: I0312 00:11:42.152842 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/caf01641-a311-4aac-af83-1edabc201169-serving-cert" (OuterVolumeSpecName: "serving-cert") pod "caf01641-a311-4aac-af83-1edabc201169" (UID: "caf01641-a311-4aac-af83-1edabc201169"). InnerVolumeSpecName "serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:11:42 crc kubenswrapper[4948]: I0312 00:11:42.153062 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/caf01641-a311-4aac-af83-1edabc201169-kube-api-access-wh6pm" (OuterVolumeSpecName: "kube-api-access-wh6pm") pod "caf01641-a311-4aac-af83-1edabc201169" (UID: "caf01641-a311-4aac-af83-1edabc201169"). InnerVolumeSpecName "kube-api-access-wh6pm". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:11:42 crc kubenswrapper[4948]: I0312 00:11:42.251570 4948 reconciler_common.go:293] "Volume detached for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/caf01641-a311-4aac-af83-1edabc201169-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 00:11:42 crc kubenswrapper[4948]: I0312 00:11:42.251615 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-wh6pm\" (UniqueName: \"kubernetes.io/projected/caf01641-a311-4aac-af83-1edabc201169-kube-api-access-wh6pm\") on node \"crc\" DevicePath \"\"" Mar 12 00:11:42 crc kubenswrapper[4948]: I0312 00:11:42.406001 4948 generic.go:334] "Generic (PLEG): container finished" podID="122cfac5-b68f-4ac2-9bd3-708c92e07c08" containerID="739f4822ffb6114ecbaef0482544cc5ee52e0f0f9d9617e822df4c63b0954128" exitCode=0 Mar 12 00:11:42 crc kubenswrapper[4948]: I0312 00:11:42.406140 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5d6f859c69-9lr2f" event={"ID":"122cfac5-b68f-4ac2-9bd3-708c92e07c08","Type":"ContainerDied","Data":"739f4822ffb6114ecbaef0482544cc5ee52e0f0f9d9617e822df4c63b0954128"} Mar 12 00:11:42 crc kubenswrapper[4948]: I0312 00:11:42.406230 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-5d6f859c69-9lr2f" event={"ID":"122cfac5-b68f-4ac2-9bd3-708c92e07c08","Type":"ContainerDied","Data":"d70fdfbf6ea660cbf04d24d833818679a0c91de9f91e6f689191d9984f783f77"} Mar 12 00:11:42 crc kubenswrapper[4948]: I0312 00:11:42.406278 4948 scope.go:117] "RemoveContainer" containerID="739f4822ffb6114ecbaef0482544cc5ee52e0f0f9d9617e822df4c63b0954128" Mar 12 00:11:42 crc kubenswrapper[4948]: I0312 00:11:42.407471 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-5d6f859c69-9lr2f" Mar 12 00:11:42 crc kubenswrapper[4948]: I0312 00:11:42.408667 4948 generic.go:334] "Generic (PLEG): container finished" podID="caf01641-a311-4aac-af83-1edabc201169" containerID="1a7cdabe724b78d08c5ce27a48568c313fa9af595a81c05ec244574a1874f459" exitCode=0 Mar 12 00:11:42 crc kubenswrapper[4948]: I0312 00:11:42.408691 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-76fbfc78f5-ll5hs" event={"ID":"caf01641-a311-4aac-af83-1edabc201169","Type":"ContainerDied","Data":"1a7cdabe724b78d08c5ce27a48568c313fa9af595a81c05ec244574a1874f459"} Mar 12 00:11:42 crc kubenswrapper[4948]: I0312 00:11:42.408717 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-76fbfc78f5-ll5hs" event={"ID":"caf01641-a311-4aac-af83-1edabc201169","Type":"ContainerDied","Data":"daa23c1151891223f66ab40d09e29fa883ebdd8edfbe1155ac83632d20cff03e"} Mar 12 00:11:42 crc kubenswrapper[4948]: I0312 00:11:42.408774 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-76fbfc78f5-ll5hs" Mar 12 00:11:42 crc kubenswrapper[4948]: I0312 00:11:42.432560 4948 scope.go:117] "RemoveContainer" containerID="739f4822ffb6114ecbaef0482544cc5ee52e0f0f9d9617e822df4c63b0954128" Mar 12 00:11:42 crc kubenswrapper[4948]: E0312 00:11:42.433984 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"739f4822ffb6114ecbaef0482544cc5ee52e0f0f9d9617e822df4c63b0954128\": container with ID starting with 739f4822ffb6114ecbaef0482544cc5ee52e0f0f9d9617e822df4c63b0954128 not found: ID does not exist" containerID="739f4822ffb6114ecbaef0482544cc5ee52e0f0f9d9617e822df4c63b0954128" Mar 12 00:11:42 crc kubenswrapper[4948]: I0312 00:11:42.434067 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"739f4822ffb6114ecbaef0482544cc5ee52e0f0f9d9617e822df4c63b0954128"} err="failed to get container status \"739f4822ffb6114ecbaef0482544cc5ee52e0f0f9d9617e822df4c63b0954128\": rpc error: code = NotFound desc = could not find container \"739f4822ffb6114ecbaef0482544cc5ee52e0f0f9d9617e822df4c63b0954128\": container with ID starting with 739f4822ffb6114ecbaef0482544cc5ee52e0f0f9d9617e822df4c63b0954128 not found: ID does not exist" Mar 12 00:11:42 crc kubenswrapper[4948]: I0312 00:11:42.434124 4948 scope.go:117] "RemoveContainer" containerID="1a7cdabe724b78d08c5ce27a48568c313fa9af595a81c05ec244574a1874f459" Mar 12 00:11:42 crc kubenswrapper[4948]: I0312 00:11:42.465622 4948 scope.go:117] "RemoveContainer" containerID="1a7cdabe724b78d08c5ce27a48568c313fa9af595a81c05ec244574a1874f459" Mar 12 00:11:42 crc kubenswrapper[4948]: E0312 00:11:42.466273 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1a7cdabe724b78d08c5ce27a48568c313fa9af595a81c05ec244574a1874f459\": container with ID starting with 1a7cdabe724b78d08c5ce27a48568c313fa9af595a81c05ec244574a1874f459 not found: ID does not exist" containerID="1a7cdabe724b78d08c5ce27a48568c313fa9af595a81c05ec244574a1874f459" Mar 12 00:11:42 crc kubenswrapper[4948]: I0312 00:11:42.466451 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1a7cdabe724b78d08c5ce27a48568c313fa9af595a81c05ec244574a1874f459"} err="failed to get container status \"1a7cdabe724b78d08c5ce27a48568c313fa9af595a81c05ec244574a1874f459\": rpc error: code = NotFound desc = could not find container \"1a7cdabe724b78d08c5ce27a48568c313fa9af595a81c05ec244574a1874f459\": container with ID starting with 1a7cdabe724b78d08c5ce27a48568c313fa9af595a81c05ec244574a1874f459 not found: ID does not exist" Mar 12 00:11:42 crc kubenswrapper[4948]: I0312 00:11:42.468118 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-controller-manager/controller-manager-76fbfc78f5-ll5hs"] Mar 12 00:11:42 crc kubenswrapper[4948]: I0312 00:11:42.485456 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-controller-manager/controller-manager-76fbfc78f5-ll5hs"] Mar 12 00:11:42 crc kubenswrapper[4948]: I0312 00:11:42.492219 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5d6f859c69-9lr2f"] Mar 12 00:11:42 crc kubenswrapper[4948]: I0312 00:11:42.499170 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-5d6f859c69-9lr2f"] Mar 12 00:11:43 crc kubenswrapper[4948]: I0312 00:11:43.290767 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-controller-manager/controller-manager-774889b746-ksff9"] Mar 12 00:11:43 crc kubenswrapper[4948]: E0312 00:11:43.291028 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="122cfac5-b68f-4ac2-9bd3-708c92e07c08" containerName="route-controller-manager" Mar 12 00:11:43 crc kubenswrapper[4948]: I0312 00:11:43.291047 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="122cfac5-b68f-4ac2-9bd3-708c92e07c08" containerName="route-controller-manager" Mar 12 00:11:43 crc kubenswrapper[4948]: E0312 00:11:43.291080 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="caf01641-a311-4aac-af83-1edabc201169" containerName="controller-manager" Mar 12 00:11:43 crc kubenswrapper[4948]: I0312 00:11:43.291091 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="caf01641-a311-4aac-af83-1edabc201169" containerName="controller-manager" Mar 12 00:11:43 crc kubenswrapper[4948]: I0312 00:11:43.291220 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="caf01641-a311-4aac-af83-1edabc201169" containerName="controller-manager" Mar 12 00:11:43 crc kubenswrapper[4948]: I0312 00:11:43.291249 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="122cfac5-b68f-4ac2-9bd3-708c92e07c08" containerName="route-controller-manager" Mar 12 00:11:43 crc kubenswrapper[4948]: I0312 00:11:43.291835 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-774889b746-ksff9" Mar 12 00:11:43 crc kubenswrapper[4948]: I0312 00:11:43.294056 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 12 00:11:43 crc kubenswrapper[4948]: I0312 00:11:43.296219 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 12 00:11:43 crc kubenswrapper[4948]: I0312 00:11:43.297069 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 12 00:11:43 crc kubenswrapper[4948]: I0312 00:11:43.297181 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-route-controller-manager/route-controller-manager-546f4f47f4-7fhfb"] Mar 12 00:11:43 crc kubenswrapper[4948]: I0312 00:11:43.297646 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 12 00:11:43 crc kubenswrapper[4948]: I0312 00:11:43.298021 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 12 00:11:43 crc kubenswrapper[4948]: I0312 00:11:43.298821 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-546f4f47f4-7fhfb" Mar 12 00:11:43 crc kubenswrapper[4948]: I0312 00:11:43.300083 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 12 00:11:43 crc kubenswrapper[4948]: I0312 00:11:43.301519 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 12 00:11:43 crc kubenswrapper[4948]: I0312 00:11:43.306062 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 12 00:11:43 crc kubenswrapper[4948]: I0312 00:11:43.306141 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 12 00:11:43 crc kubenswrapper[4948]: I0312 00:11:43.305678 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 12 00:11:43 crc kubenswrapper[4948]: I0312 00:11:43.305916 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 12 00:11:43 crc kubenswrapper[4948]: I0312 00:11:43.307014 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 12 00:11:43 crc kubenswrapper[4948]: I0312 00:11:43.309684 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 12 00:11:43 crc kubenswrapper[4948]: I0312 00:11:43.341173 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="122cfac5-b68f-4ac2-9bd3-708c92e07c08" path="/var/lib/kubelet/pods/122cfac5-b68f-4ac2-9bd3-708c92e07c08/volumes" Mar 12 00:11:43 crc kubenswrapper[4948]: I0312 00:11:43.341713 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="caf01641-a311-4aac-af83-1edabc201169" path="/var/lib/kubelet/pods/caf01641-a311-4aac-af83-1edabc201169/volumes" Mar 12 00:11:43 crc kubenswrapper[4948]: I0312 00:11:43.342079 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-546f4f47f4-7fhfb"] Mar 12 00:11:43 crc kubenswrapper[4948]: I0312 00:11:43.342110 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-774889b746-ksff9"] Mar 12 00:11:43 crc kubenswrapper[4948]: I0312 00:11:43.467906 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-5jq77\" (UniqueName: \"kubernetes.io/projected/41bc37f6-a699-43b3-801f-ad016ed06537-kube-api-access-5jq77\") pod \"controller-manager-774889b746-ksff9\" (UID: \"41bc37f6-a699-43b3-801f-ad016ed06537\") " pod="openshift-controller-manager/controller-manager-774889b746-ksff9" Mar 12 00:11:43 crc kubenswrapper[4948]: I0312 00:11:43.467972 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/16e12a46-48f4-4890-9ce6-78f26940b2dd-config\") pod \"route-controller-manager-546f4f47f4-7fhfb\" (UID: \"16e12a46-48f4-4890-9ce6-78f26940b2dd\") " pod="openshift-route-controller-manager/route-controller-manager-546f4f47f4-7fhfb" Mar 12 00:11:43 crc kubenswrapper[4948]: I0312 00:11:43.468089 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-sx4cm\" (UniqueName: \"kubernetes.io/projected/16e12a46-48f4-4890-9ce6-78f26940b2dd-kube-api-access-sx4cm\") pod \"route-controller-manager-546f4f47f4-7fhfb\" (UID: \"16e12a46-48f4-4890-9ce6-78f26940b2dd\") " pod="openshift-route-controller-manager/route-controller-manager-546f4f47f4-7fhfb" Mar 12 00:11:43 crc kubenswrapper[4948]: I0312 00:11:43.468116 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/41bc37f6-a699-43b3-801f-ad016ed06537-config\") pod \"controller-manager-774889b746-ksff9\" (UID: \"41bc37f6-a699-43b3-801f-ad016ed06537\") " pod="openshift-controller-manager/controller-manager-774889b746-ksff9" Mar 12 00:11:43 crc kubenswrapper[4948]: I0312 00:11:43.468180 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/41bc37f6-a699-43b3-801f-ad016ed06537-serving-cert\") pod \"controller-manager-774889b746-ksff9\" (UID: \"41bc37f6-a699-43b3-801f-ad016ed06537\") " pod="openshift-controller-manager/controller-manager-774889b746-ksff9" Mar 12 00:11:43 crc kubenswrapper[4948]: I0312 00:11:43.468207 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/41bc37f6-a699-43b3-801f-ad016ed06537-proxy-ca-bundles\") pod \"controller-manager-774889b746-ksff9\" (UID: \"41bc37f6-a699-43b3-801f-ad016ed06537\") " pod="openshift-controller-manager/controller-manager-774889b746-ksff9" Mar 12 00:11:43 crc kubenswrapper[4948]: I0312 00:11:43.468230 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/16e12a46-48f4-4890-9ce6-78f26940b2dd-client-ca\") pod \"route-controller-manager-546f4f47f4-7fhfb\" (UID: \"16e12a46-48f4-4890-9ce6-78f26940b2dd\") " pod="openshift-route-controller-manager/route-controller-manager-546f4f47f4-7fhfb" Mar 12 00:11:43 crc kubenswrapper[4948]: I0312 00:11:43.468252 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/41bc37f6-a699-43b3-801f-ad016ed06537-client-ca\") pod \"controller-manager-774889b746-ksff9\" (UID: \"41bc37f6-a699-43b3-801f-ad016ed06537\") " pod="openshift-controller-manager/controller-manager-774889b746-ksff9" Mar 12 00:11:43 crc kubenswrapper[4948]: I0312 00:11:43.468462 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/16e12a46-48f4-4890-9ce6-78f26940b2dd-serving-cert\") pod \"route-controller-manager-546f4f47f4-7fhfb\" (UID: \"16e12a46-48f4-4890-9ce6-78f26940b2dd\") " pod="openshift-route-controller-manager/route-controller-manager-546f4f47f4-7fhfb" Mar 12 00:11:43 crc kubenswrapper[4948]: I0312 00:11:43.570400 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-5jq77\" (UniqueName: \"kubernetes.io/projected/41bc37f6-a699-43b3-801f-ad016ed06537-kube-api-access-5jq77\") pod \"controller-manager-774889b746-ksff9\" (UID: \"41bc37f6-a699-43b3-801f-ad016ed06537\") " pod="openshift-controller-manager/controller-manager-774889b746-ksff9" Mar 12 00:11:43 crc kubenswrapper[4948]: I0312 00:11:43.570508 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/16e12a46-48f4-4890-9ce6-78f26940b2dd-config\") pod \"route-controller-manager-546f4f47f4-7fhfb\" (UID: \"16e12a46-48f4-4890-9ce6-78f26940b2dd\") " pod="openshift-route-controller-manager/route-controller-manager-546f4f47f4-7fhfb" Mar 12 00:11:43 crc kubenswrapper[4948]: I0312 00:11:43.570593 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-sx4cm\" (UniqueName: \"kubernetes.io/projected/16e12a46-48f4-4890-9ce6-78f26940b2dd-kube-api-access-sx4cm\") pod \"route-controller-manager-546f4f47f4-7fhfb\" (UID: \"16e12a46-48f4-4890-9ce6-78f26940b2dd\") " pod="openshift-route-controller-manager/route-controller-manager-546f4f47f4-7fhfb" Mar 12 00:11:43 crc kubenswrapper[4948]: I0312 00:11:43.570631 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config\" (UniqueName: \"kubernetes.io/configmap/41bc37f6-a699-43b3-801f-ad016ed06537-config\") pod \"controller-manager-774889b746-ksff9\" (UID: \"41bc37f6-a699-43b3-801f-ad016ed06537\") " pod="openshift-controller-manager/controller-manager-774889b746-ksff9" Mar 12 00:11:43 crc kubenswrapper[4948]: I0312 00:11:43.570697 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/41bc37f6-a699-43b3-801f-ad016ed06537-serving-cert\") pod \"controller-manager-774889b746-ksff9\" (UID: \"41bc37f6-a699-43b3-801f-ad016ed06537\") " pod="openshift-controller-manager/controller-manager-774889b746-ksff9" Mar 12 00:11:43 crc kubenswrapper[4948]: I0312 00:11:43.570764 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/41bc37f6-a699-43b3-801f-ad016ed06537-proxy-ca-bundles\") pod \"controller-manager-774889b746-ksff9\" (UID: \"41bc37f6-a699-43b3-801f-ad016ed06537\") " pod="openshift-controller-manager/controller-manager-774889b746-ksff9" Mar 12 00:11:43 crc kubenswrapper[4948]: I0312 00:11:43.570818 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/16e12a46-48f4-4890-9ce6-78f26940b2dd-client-ca\") pod \"route-controller-manager-546f4f47f4-7fhfb\" (UID: \"16e12a46-48f4-4890-9ce6-78f26940b2dd\") " pod="openshift-route-controller-manager/route-controller-manager-546f4f47f4-7fhfb" Mar 12 00:11:43 crc kubenswrapper[4948]: I0312 00:11:43.570868 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/41bc37f6-a699-43b3-801f-ad016ed06537-client-ca\") pod \"controller-manager-774889b746-ksff9\" (UID: \"41bc37f6-a699-43b3-801f-ad016ed06537\") " pod="openshift-controller-manager/controller-manager-774889b746-ksff9" Mar 12 00:11:43 crc kubenswrapper[4948]: I0312 00:11:43.570969 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/16e12a46-48f4-4890-9ce6-78f26940b2dd-serving-cert\") pod \"route-controller-manager-546f4f47f4-7fhfb\" (UID: \"16e12a46-48f4-4890-9ce6-78f26940b2dd\") " pod="openshift-route-controller-manager/route-controller-manager-546f4f47f4-7fhfb" Mar 12 00:11:43 crc kubenswrapper[4948]: I0312 00:11:43.572563 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"proxy-ca-bundles\" (UniqueName: \"kubernetes.io/configmap/41bc37f6-a699-43b3-801f-ad016ed06537-proxy-ca-bundles\") pod \"controller-manager-774889b746-ksff9\" (UID: \"41bc37f6-a699-43b3-801f-ad016ed06537\") " pod="openshift-controller-manager/controller-manager-774889b746-ksff9" Mar 12 00:11:43 crc kubenswrapper[4948]: I0312 00:11:43.573181 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/16e12a46-48f4-4890-9ce6-78f26940b2dd-client-ca\") pod \"route-controller-manager-546f4f47f4-7fhfb\" (UID: \"16e12a46-48f4-4890-9ce6-78f26940b2dd\") " pod="openshift-route-controller-manager/route-controller-manager-546f4f47f4-7fhfb" Mar 12 00:11:43 crc kubenswrapper[4948]: I0312 00:11:43.574076 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/41bc37f6-a699-43b3-801f-ad016ed06537-config\") pod \"controller-manager-774889b746-ksff9\" (UID: \"41bc37f6-a699-43b3-801f-ad016ed06537\") " pod="openshift-controller-manager/controller-manager-774889b746-ksff9" Mar 12 00:11:43 crc kubenswrapper[4948]: I0312 00:11:43.574637 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"client-ca\" (UniqueName: \"kubernetes.io/configmap/41bc37f6-a699-43b3-801f-ad016ed06537-client-ca\") pod \"controller-manager-774889b746-ksff9\" (UID: \"41bc37f6-a699-43b3-801f-ad016ed06537\") " pod="openshift-controller-manager/controller-manager-774889b746-ksff9" Mar 12 00:11:43 crc kubenswrapper[4948]: I0312 00:11:43.574666 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config\" (UniqueName: \"kubernetes.io/configmap/16e12a46-48f4-4890-9ce6-78f26940b2dd-config\") pod \"route-controller-manager-546f4f47f4-7fhfb\" (UID: \"16e12a46-48f4-4890-9ce6-78f26940b2dd\") " pod="openshift-route-controller-manager/route-controller-manager-546f4f47f4-7fhfb" Mar 12 00:11:43 crc kubenswrapper[4948]: I0312 00:11:43.578015 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/41bc37f6-a699-43b3-801f-ad016ed06537-serving-cert\") pod \"controller-manager-774889b746-ksff9\" (UID: \"41bc37f6-a699-43b3-801f-ad016ed06537\") " pod="openshift-controller-manager/controller-manager-774889b746-ksff9" Mar 12 00:11:43 crc kubenswrapper[4948]: I0312 00:11:43.578937 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"serving-cert\" (UniqueName: \"kubernetes.io/secret/16e12a46-48f4-4890-9ce6-78f26940b2dd-serving-cert\") pod \"route-controller-manager-546f4f47f4-7fhfb\" (UID: \"16e12a46-48f4-4890-9ce6-78f26940b2dd\") " pod="openshift-route-controller-manager/route-controller-manager-546f4f47f4-7fhfb" Mar 12 00:11:43 crc kubenswrapper[4948]: I0312 00:11:43.592795 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-sx4cm\" (UniqueName: \"kubernetes.io/projected/16e12a46-48f4-4890-9ce6-78f26940b2dd-kube-api-access-sx4cm\") pod \"route-controller-manager-546f4f47f4-7fhfb\" (UID: \"16e12a46-48f4-4890-9ce6-78f26940b2dd\") " pod="openshift-route-controller-manager/route-controller-manager-546f4f47f4-7fhfb" Mar 12 00:11:43 crc kubenswrapper[4948]: I0312 00:11:43.603528 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-5jq77\" (UniqueName: \"kubernetes.io/projected/41bc37f6-a699-43b3-801f-ad016ed06537-kube-api-access-5jq77\") pod \"controller-manager-774889b746-ksff9\" (UID: \"41bc37f6-a699-43b3-801f-ad016ed06537\") " pod="openshift-controller-manager/controller-manager-774889b746-ksff9" Mar 12 00:11:43 crc kubenswrapper[4948]: I0312 00:11:43.631468 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-controller-manager/controller-manager-774889b746-ksff9" Mar 12 00:11:43 crc kubenswrapper[4948]: I0312 00:11:43.648743 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-route-controller-manager/route-controller-manager-546f4f47f4-7fhfb" Mar 12 00:11:43 crc kubenswrapper[4948]: I0312 00:11:43.899583 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-controller-manager/controller-manager-774889b746-ksff9"] Mar 12 00:11:43 crc kubenswrapper[4948]: I0312 00:11:43.960285 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-route-controller-manager/route-controller-manager-546f4f47f4-7fhfb"] Mar 12 00:11:43 crc kubenswrapper[4948]: W0312 00:11:43.979597 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod16e12a46_48f4_4890_9ce6_78f26940b2dd.slice/crio-d7a16c9ae32fbddb2a01e43276ac9f6ce4b0dbac89e937f24960b68857de075c WatchSource:0}: Error finding container d7a16c9ae32fbddb2a01e43276ac9f6ce4b0dbac89e937f24960b68857de075c: Status 404 returned error can't find the container with id d7a16c9ae32fbddb2a01e43276ac9f6ce4b0dbac89e937f24960b68857de075c Mar 12 00:11:44 crc kubenswrapper[4948]: I0312 00:11:44.426484 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-546f4f47f4-7fhfb" event={"ID":"16e12a46-48f4-4890-9ce6-78f26940b2dd","Type":"ContainerStarted","Data":"851e8ff6f76137ffcc614fadd2e69ad7796c2fd1ca2343be60df341f115469d1"} Mar 12 00:11:44 crc kubenswrapper[4948]: I0312 00:11:44.426527 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-route-controller-manager/route-controller-manager-546f4f47f4-7fhfb" event={"ID":"16e12a46-48f4-4890-9ce6-78f26940b2dd","Type":"ContainerStarted","Data":"d7a16c9ae32fbddb2a01e43276ac9f6ce4b0dbac89e937f24960b68857de075c"} Mar 12 00:11:44 crc kubenswrapper[4948]: I0312 00:11:44.427579 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-route-controller-manager/route-controller-manager-546f4f47f4-7fhfb" Mar 12 00:11:44 crc kubenswrapper[4948]: I0312 00:11:44.428818 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-774889b746-ksff9" event={"ID":"41bc37f6-a699-43b3-801f-ad016ed06537","Type":"ContainerStarted","Data":"2bce15582c5118e04cf529907910c3e1850478fcfa95facb2b89a1c90c0c2e61"} Mar 12 00:11:44 crc kubenswrapper[4948]: I0312 00:11:44.428850 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-controller-manager/controller-manager-774889b746-ksff9" event={"ID":"41bc37f6-a699-43b3-801f-ad016ed06537","Type":"ContainerStarted","Data":"96863b00a8c3fd179c7c01e3b3a0856679d1bd515c770b7ce2fcbe20c7bcf8d8"} Mar 12 00:11:44 crc kubenswrapper[4948]: I0312 00:11:44.429227 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-controller-manager/controller-manager-774889b746-ksff9" Mar 12 00:11:44 crc kubenswrapper[4948]: I0312 00:11:44.433889 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-controller-manager/controller-manager-774889b746-ksff9" Mar 12 00:11:44 crc kubenswrapper[4948]: I0312 00:11:44.443539 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-route-controller-manager/route-controller-manager-546f4f47f4-7fhfb" podStartSLOduration=3.443526166 podStartE2EDuration="3.443526166s" podCreationTimestamp="2026-03-12 00:11:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:11:44.44244541 +0000 UTC m=+303.898049158" watchObservedRunningTime="2026-03-12 00:11:44.443526166 +0000 UTC m=+303.899129904" Mar 12 00:11:44 crc kubenswrapper[4948]: I0312 00:11:44.458030 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-controller-manager/controller-manager-774889b746-ksff9" podStartSLOduration=3.45801516 podStartE2EDuration="3.45801516s" podCreationTimestamp="2026-03-12 00:11:41 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:11:44.457126616 +0000 UTC m=+303.912730354" watchObservedRunningTime="2026-03-12 00:11:44.45801516 +0000 UTC m=+303.913618888" Mar 12 00:11:44 crc kubenswrapper[4948]: I0312 00:11:44.699855 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-route-controller-manager/route-controller-manager-546f4f47f4-7fhfb" Mar 12 00:11:46 crc kubenswrapper[4948]: I0312 00:11:46.699702 4948 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 12 00:11:46 crc kubenswrapper[4948]: I0312 00:11:46.700885 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 12 00:11:46 crc kubenswrapper[4948]: I0312 00:11:46.702388 4948 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 12 00:11:46 crc kubenswrapper[4948]: I0312 00:11:46.703951 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" containerID="cri-o://8028fd7b3e6d26bf30666425c843933b8fb87c3b2a258d510608ec718eb4fe82" gracePeriod=15 Mar 12 00:11:46 crc kubenswrapper[4948]: I0312 00:11:46.704006 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" containerID="cri-o://1ad13662849131222d5c8ba1041b81fa75bcb1fcbfa2e71c0c751f04b532a5b3" gracePeriod=15 Mar 12 00:11:46 crc kubenswrapper[4948]: I0312 00:11:46.704012 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" containerID="cri-o://1cb328c8431b39dac735d49726a2eae40a367849b2c146bb2e49d02958f0e5a2" gracePeriod=15 Mar 12 00:11:46 crc kubenswrapper[4948]: I0312 00:11:46.704029 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" containerID="cri-o://c8cd4faf47b25395ffd42d4bf5416ea29dcaa3ec380d6d6d6a9f13321d35195a" gracePeriod=15 Mar 12 00:11:46 crc kubenswrapper[4948]: I0312 00:11:46.704092 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" containerID="cri-o://d56e32f9cb96ad103bab80d6a730b70d3387264a25a7a61ef1f0726e7290d6c1" gracePeriod=15 Mar 12 00:11:46 crc kubenswrapper[4948]: I0312 00:11:46.704518 4948 kubelet.go:2421] "SyncLoop ADD" source="file" pods=["openshift-kube-apiserver/kube-apiserver-crc"] Mar 12 00:11:46 crc kubenswrapper[4948]: E0312 00:11:46.704711 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Mar 12 00:11:46 crc kubenswrapper[4948]: I0312 00:11:46.704730 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="setup" Mar 12 00:11:46 crc kubenswrapper[4948]: E0312 00:11:46.704744 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 12 00:11:46 crc kubenswrapper[4948]: I0312 00:11:46.704753 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 12 00:11:46 crc kubenswrapper[4948]: E0312 00:11:46.704766 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 12 00:11:46 crc kubenswrapper[4948]: I0312 00:11:46.704774 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 12 00:11:46 crc kubenswrapper[4948]: E0312 00:11:46.704785 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 12 00:11:46 crc kubenswrapper[4948]: I0312 00:11:46.704796 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 12 00:11:46 crc kubenswrapper[4948]: E0312 00:11:46.704815 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 12 00:11:46 crc kubenswrapper[4948]: I0312 00:11:46.704824 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 12 00:11:46 crc kubenswrapper[4948]: E0312 00:11:46.704833 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 12 00:11:46 crc kubenswrapper[4948]: I0312 00:11:46.704841 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 12 00:11:46 crc kubenswrapper[4948]: E0312 00:11:46.704855 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 12 00:11:46 crc kubenswrapper[4948]: I0312 00:11:46.704863 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 12 00:11:46 crc kubenswrapper[4948]: E0312 00:11:46.704873 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 12 00:11:46 crc kubenswrapper[4948]: I0312 00:11:46.704881 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 12 00:11:46 crc kubenswrapper[4948]: I0312 00:11:46.705018 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-insecure-readyz" Mar 12 00:11:46 crc kubenswrapper[4948]: I0312 00:11:46.705030 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 12 00:11:46 crc kubenswrapper[4948]: I0312 00:11:46.705043 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-syncer" Mar 12 00:11:46 crc kubenswrapper[4948]: I0312 00:11:46.705053 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-cert-regeneration-controller" Mar 12 00:11:46 crc kubenswrapper[4948]: I0312 00:11:46.705064 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 12 00:11:46 crc kubenswrapper[4948]: I0312 00:11:46.705075 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 12 00:11:46 crc kubenswrapper[4948]: I0312 00:11:46.705085 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver" Mar 12 00:11:46 crc kubenswrapper[4948]: I0312 00:11:46.705099 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 12 00:11:46 crc kubenswrapper[4948]: E0312 00:11:46.705213 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 12 00:11:46 crc kubenswrapper[4948]: I0312 00:11:46.705223 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 12 00:11:46 crc kubenswrapper[4948]: E0312 00:11:46.705235 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 12 00:11:46 crc kubenswrapper[4948]: I0312 00:11:46.705243 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 12 00:11:46 crc kubenswrapper[4948]: I0312 00:11:46.705400 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="f4b27818a5e8e43d0dc095d08835c792" containerName="kube-apiserver-check-endpoints" Mar 12 00:11:46 crc kubenswrapper[4948]: E0312 00:11:46.748746 4948 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.180:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 12 00:11:46 crc kubenswrapper[4948]: I0312 00:11:46.824192 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 12 00:11:46 crc kubenswrapper[4948]: I0312 00:11:46.824248 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 00:11:46 crc kubenswrapper[4948]: I0312 00:11:46.824286 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 12 00:11:46 crc kubenswrapper[4948]: I0312 00:11:46.824328 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 12 00:11:46 crc kubenswrapper[4948]: I0312 00:11:46.824351 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 00:11:46 crc kubenswrapper[4948]: I0312 00:11:46.824391 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 12 00:11:46 crc kubenswrapper[4948]: I0312 00:11:46.824418 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 12 00:11:46 crc kubenswrapper[4948]: I0312 00:11:46.828814 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 00:11:46 crc kubenswrapper[4948]: I0312 00:11:46.930638 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 12 00:11:46 crc kubenswrapper[4948]: I0312 00:11:46.930697 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 12 00:11:46 crc kubenswrapper[4948]: I0312 00:11:46.930728 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 00:11:46 crc kubenswrapper[4948]: I0312 00:11:46.930783 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 12 00:11:46 crc kubenswrapper[4948]: I0312 00:11:46.930812 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 00:11:46 crc kubenswrapper[4948]: I0312 00:11:46.930846 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 12 00:11:46 crc kubenswrapper[4948]: I0312 00:11:46.930844 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 12 00:11:46 crc kubenswrapper[4948]: I0312 00:11:46.930872 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 00:11:46 crc kubenswrapper[4948]: I0312 00:11:46.930915 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 12 00:11:46 crc kubenswrapper[4948]: I0312 00:11:46.930922 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-resource-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 00:11:46 crc kubenswrapper[4948]: I0312 00:11:46.930977 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 12 00:11:46 crc kubenswrapper[4948]: I0312 00:11:46.930983 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-audit-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 00:11:46 crc kubenswrapper[4948]: I0312 00:11:46.931009 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/71bb4a3aecc4ba5b26c4b7318770ce13-cert-dir\") pod \"kube-apiserver-crc\" (UID: \"71bb4a3aecc4ba5b26c4b7318770ce13\") " pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 00:11:46 crc kubenswrapper[4948]: I0312 00:11:46.931047 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 12 00:11:46 crc kubenswrapper[4948]: I0312 00:11:46.931048 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 12 00:11:46 crc kubenswrapper[4948]: I0312 00:11:46.931074 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"kube-apiserver-startup-monitor-crc\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 12 00:11:47 crc kubenswrapper[4948]: I0312 00:11:47.049811 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 12 00:11:47 crc kubenswrapper[4948]: W0312 00:11:47.078513 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podf85e55b1a89d02b0cb034b1ea31ed45a.slice/crio-ea459986e70760703621cef124dd1168b9492271b8f09bef31c51abaa1b57b9f WatchSource:0}: Error finding container ea459986e70760703621cef124dd1168b9492271b8f09bef31c51abaa1b57b9f: Status 404 returned error can't find the container with id ea459986e70760703621cef124dd1168b9492271b8f09bef31c51abaa1b57b9f Mar 12 00:11:47 crc kubenswrapper[4948]: E0312 00:11:47.081950 4948 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.180:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189bef964aea31e5 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:11:47.081417189 +0000 UTC m=+306.537020937,LastTimestamp:2026-03-12 00:11:47.081417189 +0000 UTC m=+306.537020937,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:11:47 crc kubenswrapper[4948]: I0312 00:11:47.448367 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"b8f42af8d5706c672759fe6b3ea65eaccb27bc3e80ed254e63660d83b69bfcde"} Mar 12 00:11:47 crc kubenswrapper[4948]: I0312 00:11:47.448442 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" event={"ID":"f85e55b1a89d02b0cb034b1ea31ed45a","Type":"ContainerStarted","Data":"ea459986e70760703621cef124dd1168b9492271b8f09bef31c51abaa1b57b9f"} Mar 12 00:11:47 crc kubenswrapper[4948]: E0312 00:11:47.449780 4948 kubelet.go:1929] "Failed creating a mirror pod for" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods\": dial tcp 38.102.83.180:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 12 00:11:47 crc kubenswrapper[4948]: I0312 00:11:47.451203 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-check-endpoints/3.log" Mar 12 00:11:47 crc kubenswrapper[4948]: I0312 00:11:47.453227 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 12 00:11:47 crc kubenswrapper[4948]: I0312 00:11:47.454254 4948 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="8028fd7b3e6d26bf30666425c843933b8fb87c3b2a258d510608ec718eb4fe82" exitCode=0 Mar 12 00:11:47 crc kubenswrapper[4948]: I0312 00:11:47.454293 4948 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="d56e32f9cb96ad103bab80d6a730b70d3387264a25a7a61ef1f0726e7290d6c1" exitCode=0 Mar 12 00:11:47 crc kubenswrapper[4948]: I0312 00:11:47.454346 4948 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="1ad13662849131222d5c8ba1041b81fa75bcb1fcbfa2e71c0c751f04b532a5b3" exitCode=0 Mar 12 00:11:47 crc kubenswrapper[4948]: I0312 00:11:47.454356 4948 scope.go:117] "RemoveContainer" containerID="73834c8db23f6536034dce09e38b56665211238280bf9a4e1b13e74d81245a77" Mar 12 00:11:47 crc kubenswrapper[4948]: I0312 00:11:47.454374 4948 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="c8cd4faf47b25395ffd42d4bf5416ea29dcaa3ec380d6d6d6a9f13321d35195a" exitCode=2 Mar 12 00:11:47 crc kubenswrapper[4948]: I0312 00:11:47.456767 4948 generic.go:334] "Generic (PLEG): container finished" podID="62c7736b-4d37-43bf-8ee8-f1e5c3ac3a37" containerID="134eb20d8890c76b8433a37aaa725ef8f2322110a3d18539d34370335a2841f3" exitCode=0 Mar 12 00:11:47 crc kubenswrapper[4948]: I0312 00:11:47.456818 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"62c7736b-4d37-43bf-8ee8-f1e5c3ac3a37","Type":"ContainerDied","Data":"134eb20d8890c76b8433a37aaa725ef8f2322110a3d18539d34370335a2841f3"} Mar 12 00:11:47 crc kubenswrapper[4948]: I0312 00:11:47.457450 4948 status_manager.go:851] "Failed to get status for pod" podUID="62c7736b-4d37-43bf-8ee8-f1e5c3ac3a37" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Mar 12 00:11:48 crc kubenswrapper[4948]: I0312 00:11:48.472498 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 12 00:11:48 crc kubenswrapper[4948]: I0312 00:11:48.898786 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 12 00:11:48 crc kubenswrapper[4948]: I0312 00:11:48.900220 4948 status_manager.go:851] "Failed to get status for pod" podUID="62c7736b-4d37-43bf-8ee8-f1e5c3ac3a37" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Mar 12 00:11:49 crc kubenswrapper[4948]: I0312 00:11:49.061028 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/62c7736b-4d37-43bf-8ee8-f1e5c3ac3a37-var-lock\") pod \"62c7736b-4d37-43bf-8ee8-f1e5c3ac3a37\" (UID: \"62c7736b-4d37-43bf-8ee8-f1e5c3ac3a37\") " Mar 12 00:11:49 crc kubenswrapper[4948]: I0312 00:11:49.061146 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/62c7736b-4d37-43bf-8ee8-f1e5c3ac3a37-kube-api-access\") pod \"62c7736b-4d37-43bf-8ee8-f1e5c3ac3a37\" (UID: \"62c7736b-4d37-43bf-8ee8-f1e5c3ac3a37\") " Mar 12 00:11:49 crc kubenswrapper[4948]: I0312 00:11:49.061198 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/62c7736b-4d37-43bf-8ee8-f1e5c3ac3a37-kubelet-dir\") pod \"62c7736b-4d37-43bf-8ee8-f1e5c3ac3a37\" (UID: \"62c7736b-4d37-43bf-8ee8-f1e5c3ac3a37\") " Mar 12 00:11:49 crc kubenswrapper[4948]: I0312 00:11:49.061447 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/62c7736b-4d37-43bf-8ee8-f1e5c3ac3a37-kubelet-dir" (OuterVolumeSpecName: "kubelet-dir") pod "62c7736b-4d37-43bf-8ee8-f1e5c3ac3a37" (UID: "62c7736b-4d37-43bf-8ee8-f1e5c3ac3a37"). InnerVolumeSpecName "kubelet-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 00:11:49 crc kubenswrapper[4948]: I0312 00:11:49.062540 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/62c7736b-4d37-43bf-8ee8-f1e5c3ac3a37-var-lock" (OuterVolumeSpecName: "var-lock") pod "62c7736b-4d37-43bf-8ee8-f1e5c3ac3a37" (UID: "62c7736b-4d37-43bf-8ee8-f1e5c3ac3a37"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 00:11:49 crc kubenswrapper[4948]: I0312 00:11:49.068501 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/62c7736b-4d37-43bf-8ee8-f1e5c3ac3a37-kube-api-access" (OuterVolumeSpecName: "kube-api-access") pod "62c7736b-4d37-43bf-8ee8-f1e5c3ac3a37" (UID: "62c7736b-4d37-43bf-8ee8-f1e5c3ac3a37"). InnerVolumeSpecName "kube-api-access". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:11:49 crc kubenswrapper[4948]: I0312 00:11:49.078826 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 12 00:11:49 crc kubenswrapper[4948]: I0312 00:11:49.080176 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 00:11:49 crc kubenswrapper[4948]: I0312 00:11:49.080868 4948 status_manager.go:851] "Failed to get status for pod" podUID="62c7736b-4d37-43bf-8ee8-f1e5c3ac3a37" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Mar 12 00:11:49 crc kubenswrapper[4948]: I0312 00:11:49.081449 4948 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Mar 12 00:11:49 crc kubenswrapper[4948]: I0312 00:11:49.163106 4948 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/62c7736b-4d37-43bf-8ee8-f1e5c3ac3a37-var-lock\") on node \"crc\" DevicePath \"\"" Mar 12 00:11:49 crc kubenswrapper[4948]: I0312 00:11:49.163144 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access\" (UniqueName: \"kubernetes.io/projected/62c7736b-4d37-43bf-8ee8-f1e5c3ac3a37-kube-api-access\") on node \"crc\" DevicePath \"\"" Mar 12 00:11:49 crc kubenswrapper[4948]: I0312 00:11:49.163156 4948 reconciler_common.go:293] "Volume detached for volume \"kubelet-dir\" (UniqueName: \"kubernetes.io/host-path/62c7736b-4d37-43bf-8ee8-f1e5c3ac3a37-kubelet-dir\") on node \"crc\" DevicePath \"\"" Mar 12 00:11:49 crc kubenswrapper[4948]: I0312 00:11:49.264625 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 12 00:11:49 crc kubenswrapper[4948]: I0312 00:11:49.264727 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir" (OuterVolumeSpecName: "cert-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "cert-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 00:11:49 crc kubenswrapper[4948]: I0312 00:11:49.264781 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 12 00:11:49 crc kubenswrapper[4948]: I0312 00:11:49.264825 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") pod \"f4b27818a5e8e43d0dc095d08835c792\" (UID: \"f4b27818a5e8e43d0dc095d08835c792\") " Mar 12 00:11:49 crc kubenswrapper[4948]: I0312 00:11:49.264901 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 00:11:49 crc kubenswrapper[4948]: I0312 00:11:49.265032 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f4b27818a5e8e43d0dc095d08835c792" (UID: "f4b27818a5e8e43d0dc095d08835c792"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 00:11:49 crc kubenswrapper[4948]: I0312 00:11:49.265355 4948 reconciler_common.go:293] "Volume detached for volume \"cert-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-cert-dir\") on node \"crc\" DevicePath \"\"" Mar 12 00:11:49 crc kubenswrapper[4948]: I0312 00:11:49.265381 4948 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-audit-dir\") on node \"crc\" DevicePath \"\"" Mar 12 00:11:49 crc kubenswrapper[4948]: I0312 00:11:49.265399 4948 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f4b27818a5e8e43d0dc095d08835c792-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 12 00:11:49 crc kubenswrapper[4948]: I0312 00:11:49.322892 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f4b27818a5e8e43d0dc095d08835c792" path="/var/lib/kubelet/pods/f4b27818a5e8e43d0dc095d08835c792/volumes" Mar 12 00:11:49 crc kubenswrapper[4948]: I0312 00:11:49.484136 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/installer-9-crc" event={"ID":"62c7736b-4d37-43bf-8ee8-f1e5c3ac3a37","Type":"ContainerDied","Data":"9b56865be8299fc5e275b1124ac672af03fb948b0c3ddddb03d781cef1a45df3"} Mar 12 00:11:49 crc kubenswrapper[4948]: I0312 00:11:49.484225 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="9b56865be8299fc5e275b1124ac672af03fb948b0c3ddddb03d781cef1a45df3" Mar 12 00:11:49 crc kubenswrapper[4948]: I0312 00:11:49.484179 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/installer-9-crc" Mar 12 00:11:49 crc kubenswrapper[4948]: I0312 00:11:49.488637 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-crc_f4b27818a5e8e43d0dc095d08835c792/kube-apiserver-cert-syncer/0.log" Mar 12 00:11:49 crc kubenswrapper[4948]: I0312 00:11:49.489712 4948 generic.go:334] "Generic (PLEG): container finished" podID="f4b27818a5e8e43d0dc095d08835c792" containerID="1cb328c8431b39dac735d49726a2eae40a367849b2c146bb2e49d02958f0e5a2" exitCode=0 Mar 12 00:11:49 crc kubenswrapper[4948]: I0312 00:11:49.489796 4948 scope.go:117] "RemoveContainer" containerID="8028fd7b3e6d26bf30666425c843933b8fb87c3b2a258d510608ec718eb4fe82" Mar 12 00:11:49 crc kubenswrapper[4948]: I0312 00:11:49.489861 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 00:11:49 crc kubenswrapper[4948]: I0312 00:11:49.490877 4948 status_manager.go:851] "Failed to get status for pod" podUID="62c7736b-4d37-43bf-8ee8-f1e5c3ac3a37" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Mar 12 00:11:49 crc kubenswrapper[4948]: I0312 00:11:49.491510 4948 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Mar 12 00:11:49 crc kubenswrapper[4948]: I0312 00:11:49.492167 4948 status_manager.go:851] "Failed to get status for pod" podUID="62c7736b-4d37-43bf-8ee8-f1e5c3ac3a37" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Mar 12 00:11:49 crc kubenswrapper[4948]: I0312 00:11:49.492932 4948 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Mar 12 00:11:49 crc kubenswrapper[4948]: I0312 00:11:49.494533 4948 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Mar 12 00:11:49 crc kubenswrapper[4948]: I0312 00:11:49.495161 4948 status_manager.go:851] "Failed to get status for pod" podUID="62c7736b-4d37-43bf-8ee8-f1e5c3ac3a37" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Mar 12 00:11:49 crc kubenswrapper[4948]: I0312 00:11:49.511928 4948 scope.go:117] "RemoveContainer" containerID="d56e32f9cb96ad103bab80d6a730b70d3387264a25a7a61ef1f0726e7290d6c1" Mar 12 00:11:49 crc kubenswrapper[4948]: I0312 00:11:49.531905 4948 scope.go:117] "RemoveContainer" containerID="1ad13662849131222d5c8ba1041b81fa75bcb1fcbfa2e71c0c751f04b532a5b3" Mar 12 00:11:49 crc kubenswrapper[4948]: I0312 00:11:49.558878 4948 scope.go:117] "RemoveContainer" containerID="c8cd4faf47b25395ffd42d4bf5416ea29dcaa3ec380d6d6d6a9f13321d35195a" Mar 12 00:11:49 crc kubenswrapper[4948]: I0312 00:11:49.583521 4948 scope.go:117] "RemoveContainer" containerID="1cb328c8431b39dac735d49726a2eae40a367849b2c146bb2e49d02958f0e5a2" Mar 12 00:11:49 crc kubenswrapper[4948]: I0312 00:11:49.623598 4948 scope.go:117] "RemoveContainer" containerID="7e42f2e94463ac8a3d7980a25122062da9ee87fcdc1f10e260ff9974b2e42359" Mar 12 00:11:49 crc kubenswrapper[4948]: I0312 00:11:49.664415 4948 scope.go:117] "RemoveContainer" containerID="8028fd7b3e6d26bf30666425c843933b8fb87c3b2a258d510608ec718eb4fe82" Mar 12 00:11:49 crc kubenswrapper[4948]: E0312 00:11:49.665675 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8028fd7b3e6d26bf30666425c843933b8fb87c3b2a258d510608ec718eb4fe82\": container with ID starting with 8028fd7b3e6d26bf30666425c843933b8fb87c3b2a258d510608ec718eb4fe82 not found: ID does not exist" containerID="8028fd7b3e6d26bf30666425c843933b8fb87c3b2a258d510608ec718eb4fe82" Mar 12 00:11:49 crc kubenswrapper[4948]: I0312 00:11:49.665743 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8028fd7b3e6d26bf30666425c843933b8fb87c3b2a258d510608ec718eb4fe82"} err="failed to get container status \"8028fd7b3e6d26bf30666425c843933b8fb87c3b2a258d510608ec718eb4fe82\": rpc error: code = NotFound desc = could not find container \"8028fd7b3e6d26bf30666425c843933b8fb87c3b2a258d510608ec718eb4fe82\": container with ID starting with 8028fd7b3e6d26bf30666425c843933b8fb87c3b2a258d510608ec718eb4fe82 not found: ID does not exist" Mar 12 00:11:49 crc kubenswrapper[4948]: I0312 00:11:49.665787 4948 scope.go:117] "RemoveContainer" containerID="d56e32f9cb96ad103bab80d6a730b70d3387264a25a7a61ef1f0726e7290d6c1" Mar 12 00:11:49 crc kubenswrapper[4948]: E0312 00:11:49.666524 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d56e32f9cb96ad103bab80d6a730b70d3387264a25a7a61ef1f0726e7290d6c1\": container with ID starting with d56e32f9cb96ad103bab80d6a730b70d3387264a25a7a61ef1f0726e7290d6c1 not found: ID does not exist" containerID="d56e32f9cb96ad103bab80d6a730b70d3387264a25a7a61ef1f0726e7290d6c1" Mar 12 00:11:49 crc kubenswrapper[4948]: I0312 00:11:49.666579 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d56e32f9cb96ad103bab80d6a730b70d3387264a25a7a61ef1f0726e7290d6c1"} err="failed to get container status \"d56e32f9cb96ad103bab80d6a730b70d3387264a25a7a61ef1f0726e7290d6c1\": rpc error: code = NotFound desc = could not find container \"d56e32f9cb96ad103bab80d6a730b70d3387264a25a7a61ef1f0726e7290d6c1\": container with ID starting with d56e32f9cb96ad103bab80d6a730b70d3387264a25a7a61ef1f0726e7290d6c1 not found: ID does not exist" Mar 12 00:11:49 crc kubenswrapper[4948]: I0312 00:11:49.666618 4948 scope.go:117] "RemoveContainer" containerID="1ad13662849131222d5c8ba1041b81fa75bcb1fcbfa2e71c0c751f04b532a5b3" Mar 12 00:11:49 crc kubenswrapper[4948]: E0312 00:11:49.667003 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1ad13662849131222d5c8ba1041b81fa75bcb1fcbfa2e71c0c751f04b532a5b3\": container with ID starting with 1ad13662849131222d5c8ba1041b81fa75bcb1fcbfa2e71c0c751f04b532a5b3 not found: ID does not exist" containerID="1ad13662849131222d5c8ba1041b81fa75bcb1fcbfa2e71c0c751f04b532a5b3" Mar 12 00:11:49 crc kubenswrapper[4948]: I0312 00:11:49.667048 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1ad13662849131222d5c8ba1041b81fa75bcb1fcbfa2e71c0c751f04b532a5b3"} err="failed to get container status \"1ad13662849131222d5c8ba1041b81fa75bcb1fcbfa2e71c0c751f04b532a5b3\": rpc error: code = NotFound desc = could not find container \"1ad13662849131222d5c8ba1041b81fa75bcb1fcbfa2e71c0c751f04b532a5b3\": container with ID starting with 1ad13662849131222d5c8ba1041b81fa75bcb1fcbfa2e71c0c751f04b532a5b3 not found: ID does not exist" Mar 12 00:11:49 crc kubenswrapper[4948]: I0312 00:11:49.667079 4948 scope.go:117] "RemoveContainer" containerID="c8cd4faf47b25395ffd42d4bf5416ea29dcaa3ec380d6d6d6a9f13321d35195a" Mar 12 00:11:49 crc kubenswrapper[4948]: E0312 00:11:49.667706 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c8cd4faf47b25395ffd42d4bf5416ea29dcaa3ec380d6d6d6a9f13321d35195a\": container with ID starting with c8cd4faf47b25395ffd42d4bf5416ea29dcaa3ec380d6d6d6a9f13321d35195a not found: ID does not exist" containerID="c8cd4faf47b25395ffd42d4bf5416ea29dcaa3ec380d6d6d6a9f13321d35195a" Mar 12 00:11:49 crc kubenswrapper[4948]: I0312 00:11:49.667786 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c8cd4faf47b25395ffd42d4bf5416ea29dcaa3ec380d6d6d6a9f13321d35195a"} err="failed to get container status \"c8cd4faf47b25395ffd42d4bf5416ea29dcaa3ec380d6d6d6a9f13321d35195a\": rpc error: code = NotFound desc = could not find container \"c8cd4faf47b25395ffd42d4bf5416ea29dcaa3ec380d6d6d6a9f13321d35195a\": container with ID starting with c8cd4faf47b25395ffd42d4bf5416ea29dcaa3ec380d6d6d6a9f13321d35195a not found: ID does not exist" Mar 12 00:11:49 crc kubenswrapper[4948]: I0312 00:11:49.667834 4948 scope.go:117] "RemoveContainer" containerID="1cb328c8431b39dac735d49726a2eae40a367849b2c146bb2e49d02958f0e5a2" Mar 12 00:11:49 crc kubenswrapper[4948]: E0312 00:11:49.668573 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1cb328c8431b39dac735d49726a2eae40a367849b2c146bb2e49d02958f0e5a2\": container with ID starting with 1cb328c8431b39dac735d49726a2eae40a367849b2c146bb2e49d02958f0e5a2 not found: ID does not exist" containerID="1cb328c8431b39dac735d49726a2eae40a367849b2c146bb2e49d02958f0e5a2" Mar 12 00:11:49 crc kubenswrapper[4948]: I0312 00:11:49.668627 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1cb328c8431b39dac735d49726a2eae40a367849b2c146bb2e49d02958f0e5a2"} err="failed to get container status \"1cb328c8431b39dac735d49726a2eae40a367849b2c146bb2e49d02958f0e5a2\": rpc error: code = NotFound desc = could not find container \"1cb328c8431b39dac735d49726a2eae40a367849b2c146bb2e49d02958f0e5a2\": container with ID starting with 1cb328c8431b39dac735d49726a2eae40a367849b2c146bb2e49d02958f0e5a2 not found: ID does not exist" Mar 12 00:11:49 crc kubenswrapper[4948]: I0312 00:11:49.668657 4948 scope.go:117] "RemoveContainer" containerID="7e42f2e94463ac8a3d7980a25122062da9ee87fcdc1f10e260ff9974b2e42359" Mar 12 00:11:49 crc kubenswrapper[4948]: E0312 00:11:49.669171 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"7e42f2e94463ac8a3d7980a25122062da9ee87fcdc1f10e260ff9974b2e42359\": container with ID starting with 7e42f2e94463ac8a3d7980a25122062da9ee87fcdc1f10e260ff9974b2e42359 not found: ID does not exist" containerID="7e42f2e94463ac8a3d7980a25122062da9ee87fcdc1f10e260ff9974b2e42359" Mar 12 00:11:49 crc kubenswrapper[4948]: I0312 00:11:49.669231 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"7e42f2e94463ac8a3d7980a25122062da9ee87fcdc1f10e260ff9974b2e42359"} err="failed to get container status \"7e42f2e94463ac8a3d7980a25122062da9ee87fcdc1f10e260ff9974b2e42359\": rpc error: code = NotFound desc = could not find container \"7e42f2e94463ac8a3d7980a25122062da9ee87fcdc1f10e260ff9974b2e42359\": container with ID starting with 7e42f2e94463ac8a3d7980a25122062da9ee87fcdc1f10e260ff9974b2e42359 not found: ID does not exist" Mar 12 00:11:51 crc kubenswrapper[4948]: I0312 00:11:51.317988 4948 status_manager.go:851] "Failed to get status for pod" podUID="62c7736b-4d37-43bf-8ee8-f1e5c3ac3a37" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Mar 12 00:11:51 crc kubenswrapper[4948]: I0312 00:11:51.319246 4948 status_manager.go:851] "Failed to get status for pod" podUID="f4b27818a5e8e43d0dc095d08835c792" pod="openshift-kube-apiserver/kube-apiserver-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Mar 12 00:11:51 crc kubenswrapper[4948]: I0312 00:11:51.640453 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-authentication/oauth-openshift-558db77b4-qsf8x" podUID="f90934d6-9305-4513-8d84-cd6a6fcc4a59" containerName="oauth-openshift" containerID="cri-o://56f73e2a84df78402185d6ddeb2c483aaed0eb93cd03771074df5b28ef703b8a" gracePeriod=15 Mar 12 00:11:52 crc kubenswrapper[4948]: I0312 00:11:52.186830 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-qsf8x" Mar 12 00:11:52 crc kubenswrapper[4948]: I0312 00:11:52.187755 4948 status_manager.go:851] "Failed to get status for pod" podUID="f90934d6-9305-4513-8d84-cd6a6fcc4a59" pod="openshift-authentication/oauth-openshift-558db77b4-qsf8x" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-qsf8x\": dial tcp 38.102.83.180:6443: connect: connection refused" Mar 12 00:11:52 crc kubenswrapper[4948]: I0312 00:11:52.188319 4948 status_manager.go:851] "Failed to get status for pod" podUID="62c7736b-4d37-43bf-8ee8-f1e5c3ac3a37" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Mar 12 00:11:52 crc kubenswrapper[4948]: I0312 00:11:52.304811 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f90934d6-9305-4513-8d84-cd6a6fcc4a59-v4-0-config-user-template-login\") pod \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\" (UID: \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\") " Mar 12 00:11:52 crc kubenswrapper[4948]: I0312 00:11:52.304906 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f90934d6-9305-4513-8d84-cd6a6fcc4a59-v4-0-config-system-trusted-ca-bundle\") pod \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\" (UID: \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\") " Mar 12 00:11:52 crc kubenswrapper[4948]: I0312 00:11:52.304970 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f90934d6-9305-4513-8d84-cd6a6fcc4a59-v4-0-config-system-session\") pod \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\" (UID: \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\") " Mar 12 00:11:52 crc kubenswrapper[4948]: I0312 00:11:52.305058 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f90934d6-9305-4513-8d84-cd6a6fcc4a59-audit-policies\") pod \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\" (UID: \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\") " Mar 12 00:11:52 crc kubenswrapper[4948]: I0312 00:11:52.305132 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f90934d6-9305-4513-8d84-cd6a6fcc4a59-audit-dir\") pod \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\" (UID: \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\") " Mar 12 00:11:52 crc kubenswrapper[4948]: I0312 00:11:52.305182 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f90934d6-9305-4513-8d84-cd6a6fcc4a59-v4-0-config-user-idp-0-file-data\") pod \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\" (UID: \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\") " Mar 12 00:11:52 crc kubenswrapper[4948]: I0312 00:11:52.305235 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f90934d6-9305-4513-8d84-cd6a6fcc4a59-v4-0-config-system-cliconfig\") pod \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\" (UID: \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\") " Mar 12 00:11:52 crc kubenswrapper[4948]: I0312 00:11:52.305371 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f90934d6-9305-4513-8d84-cd6a6fcc4a59-v4-0-config-system-serving-cert\") pod \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\" (UID: \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\") " Mar 12 00:11:52 crc kubenswrapper[4948]: I0312 00:11:52.305378 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f90934d6-9305-4513-8d84-cd6a6fcc4a59-audit-dir" (OuterVolumeSpecName: "audit-dir") pod "f90934d6-9305-4513-8d84-cd6a6fcc4a59" (UID: "f90934d6-9305-4513-8d84-cd6a6fcc4a59"). InnerVolumeSpecName "audit-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 00:11:52 crc kubenswrapper[4948]: I0312 00:11:52.305453 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f90934d6-9305-4513-8d84-cd6a6fcc4a59-v4-0-config-system-ocp-branding-template\") pod \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\" (UID: \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\") " Mar 12 00:11:52 crc kubenswrapper[4948]: I0312 00:11:52.305512 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f90934d6-9305-4513-8d84-cd6a6fcc4a59-v4-0-config-user-template-error\") pod \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\" (UID: \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\") " Mar 12 00:11:52 crc kubenswrapper[4948]: I0312 00:11:52.305561 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f90934d6-9305-4513-8d84-cd6a6fcc4a59-v4-0-config-system-router-certs\") pod \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\" (UID: \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\") " Mar 12 00:11:52 crc kubenswrapper[4948]: I0312 00:11:52.305609 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f90934d6-9305-4513-8d84-cd6a6fcc4a59-v4-0-config-system-service-ca\") pod \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\" (UID: \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\") " Mar 12 00:11:52 crc kubenswrapper[4948]: I0312 00:11:52.305676 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dgm4m\" (UniqueName: \"kubernetes.io/projected/f90934d6-9305-4513-8d84-cd6a6fcc4a59-kube-api-access-dgm4m\") pod \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\" (UID: \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\") " Mar 12 00:11:52 crc kubenswrapper[4948]: I0312 00:11:52.305725 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f90934d6-9305-4513-8d84-cd6a6fcc4a59-v4-0-config-user-template-provider-selection\") pod \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\" (UID: \"f90934d6-9305-4513-8d84-cd6a6fcc4a59\") " Mar 12 00:11:52 crc kubenswrapper[4948]: I0312 00:11:52.306110 4948 reconciler_common.go:293] "Volume detached for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/f90934d6-9305-4513-8d84-cd6a6fcc4a59-audit-dir\") on node \"crc\" DevicePath \"\"" Mar 12 00:11:52 crc kubenswrapper[4948]: I0312 00:11:52.306557 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f90934d6-9305-4513-8d84-cd6a6fcc4a59-v4-0-config-system-cliconfig" (OuterVolumeSpecName: "v4-0-config-system-cliconfig") pod "f90934d6-9305-4513-8d84-cd6a6fcc4a59" (UID: "f90934d6-9305-4513-8d84-cd6a6fcc4a59"). InnerVolumeSpecName "v4-0-config-system-cliconfig". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:11:52 crc kubenswrapper[4948]: I0312 00:11:52.306788 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f90934d6-9305-4513-8d84-cd6a6fcc4a59-v4-0-config-system-service-ca" (OuterVolumeSpecName: "v4-0-config-system-service-ca") pod "f90934d6-9305-4513-8d84-cd6a6fcc4a59" (UID: "f90934d6-9305-4513-8d84-cd6a6fcc4a59"). InnerVolumeSpecName "v4-0-config-system-service-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:11:52 crc kubenswrapper[4948]: I0312 00:11:52.306974 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f90934d6-9305-4513-8d84-cd6a6fcc4a59-v4-0-config-system-trusted-ca-bundle" (OuterVolumeSpecName: "v4-0-config-system-trusted-ca-bundle") pod "f90934d6-9305-4513-8d84-cd6a6fcc4a59" (UID: "f90934d6-9305-4513-8d84-cd6a6fcc4a59"). InnerVolumeSpecName "v4-0-config-system-trusted-ca-bundle". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:11:52 crc kubenswrapper[4948]: I0312 00:11:52.308023 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/f90934d6-9305-4513-8d84-cd6a6fcc4a59-audit-policies" (OuterVolumeSpecName: "audit-policies") pod "f90934d6-9305-4513-8d84-cd6a6fcc4a59" (UID: "f90934d6-9305-4513-8d84-cd6a6fcc4a59"). InnerVolumeSpecName "audit-policies". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:11:52 crc kubenswrapper[4948]: I0312 00:11:52.310536 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f90934d6-9305-4513-8d84-cd6a6fcc4a59-v4-0-config-user-idp-0-file-data" (OuterVolumeSpecName: "v4-0-config-user-idp-0-file-data") pod "f90934d6-9305-4513-8d84-cd6a6fcc4a59" (UID: "f90934d6-9305-4513-8d84-cd6a6fcc4a59"). InnerVolumeSpecName "v4-0-config-user-idp-0-file-data". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:11:52 crc kubenswrapper[4948]: I0312 00:11:52.311137 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f90934d6-9305-4513-8d84-cd6a6fcc4a59-v4-0-config-system-router-certs" (OuterVolumeSpecName: "v4-0-config-system-router-certs") pod "f90934d6-9305-4513-8d84-cd6a6fcc4a59" (UID: "f90934d6-9305-4513-8d84-cd6a6fcc4a59"). InnerVolumeSpecName "v4-0-config-system-router-certs". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:11:52 crc kubenswrapper[4948]: I0312 00:11:52.311854 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f90934d6-9305-4513-8d84-cd6a6fcc4a59-v4-0-config-system-session" (OuterVolumeSpecName: "v4-0-config-system-session") pod "f90934d6-9305-4513-8d84-cd6a6fcc4a59" (UID: "f90934d6-9305-4513-8d84-cd6a6fcc4a59"). InnerVolumeSpecName "v4-0-config-system-session". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:11:52 crc kubenswrapper[4948]: I0312 00:11:52.312192 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f90934d6-9305-4513-8d84-cd6a6fcc4a59-v4-0-config-system-serving-cert" (OuterVolumeSpecName: "v4-0-config-system-serving-cert") pod "f90934d6-9305-4513-8d84-cd6a6fcc4a59" (UID: "f90934d6-9305-4513-8d84-cd6a6fcc4a59"). InnerVolumeSpecName "v4-0-config-system-serving-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:11:52 crc kubenswrapper[4948]: I0312 00:11:52.312490 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f90934d6-9305-4513-8d84-cd6a6fcc4a59-v4-0-config-user-template-error" (OuterVolumeSpecName: "v4-0-config-user-template-error") pod "f90934d6-9305-4513-8d84-cd6a6fcc4a59" (UID: "f90934d6-9305-4513-8d84-cd6a6fcc4a59"). InnerVolumeSpecName "v4-0-config-user-template-error". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:11:52 crc kubenswrapper[4948]: I0312 00:11:52.313059 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f90934d6-9305-4513-8d84-cd6a6fcc4a59-v4-0-config-system-ocp-branding-template" (OuterVolumeSpecName: "v4-0-config-system-ocp-branding-template") pod "f90934d6-9305-4513-8d84-cd6a6fcc4a59" (UID: "f90934d6-9305-4513-8d84-cd6a6fcc4a59"). InnerVolumeSpecName "v4-0-config-system-ocp-branding-template". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:11:52 crc kubenswrapper[4948]: I0312 00:11:52.314952 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f90934d6-9305-4513-8d84-cd6a6fcc4a59-v4-0-config-user-template-login" (OuterVolumeSpecName: "v4-0-config-user-template-login") pod "f90934d6-9305-4513-8d84-cd6a6fcc4a59" (UID: "f90934d6-9305-4513-8d84-cd6a6fcc4a59"). InnerVolumeSpecName "v4-0-config-user-template-login". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:11:52 crc kubenswrapper[4948]: I0312 00:11:52.321601 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/f90934d6-9305-4513-8d84-cd6a6fcc4a59-v4-0-config-user-template-provider-selection" (OuterVolumeSpecName: "v4-0-config-user-template-provider-selection") pod "f90934d6-9305-4513-8d84-cd6a6fcc4a59" (UID: "f90934d6-9305-4513-8d84-cd6a6fcc4a59"). InnerVolumeSpecName "v4-0-config-user-template-provider-selection". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:11:52 crc kubenswrapper[4948]: I0312 00:11:52.323608 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f90934d6-9305-4513-8d84-cd6a6fcc4a59-kube-api-access-dgm4m" (OuterVolumeSpecName: "kube-api-access-dgm4m") pod "f90934d6-9305-4513-8d84-cd6a6fcc4a59" (UID: "f90934d6-9305-4513-8d84-cd6a6fcc4a59"). InnerVolumeSpecName "kube-api-access-dgm4m". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:11:52 crc kubenswrapper[4948]: I0312 00:11:52.407507 4948 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/f90934d6-9305-4513-8d84-cd6a6fcc4a59-v4-0-config-user-template-provider-selection\") on node \"crc\" DevicePath \"\"" Mar 12 00:11:52 crc kubenswrapper[4948]: I0312 00:11:52.408704 4948 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/f90934d6-9305-4513-8d84-cd6a6fcc4a59-v4-0-config-user-template-login\") on node \"crc\" DevicePath \"\"" Mar 12 00:11:52 crc kubenswrapper[4948]: I0312 00:11:52.408748 4948 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/f90934d6-9305-4513-8d84-cd6a6fcc4a59-v4-0-config-system-trusted-ca-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 00:11:52 crc kubenswrapper[4948]: I0312 00:11:52.408770 4948 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/f90934d6-9305-4513-8d84-cd6a6fcc4a59-v4-0-config-system-session\") on node \"crc\" DevicePath \"\"" Mar 12 00:11:52 crc kubenswrapper[4948]: I0312 00:11:52.408792 4948 reconciler_common.go:293] "Volume detached for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/f90934d6-9305-4513-8d84-cd6a6fcc4a59-audit-policies\") on node \"crc\" DevicePath \"\"" Mar 12 00:11:52 crc kubenswrapper[4948]: I0312 00:11:52.408812 4948 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/f90934d6-9305-4513-8d84-cd6a6fcc4a59-v4-0-config-user-idp-0-file-data\") on node \"crc\" DevicePath \"\"" Mar 12 00:11:52 crc kubenswrapper[4948]: I0312 00:11:52.408832 4948 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/f90934d6-9305-4513-8d84-cd6a6fcc4a59-v4-0-config-system-cliconfig\") on node \"crc\" DevicePath \"\"" Mar 12 00:11:52 crc kubenswrapper[4948]: I0312 00:11:52.408851 4948 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/f90934d6-9305-4513-8d84-cd6a6fcc4a59-v4-0-config-system-serving-cert\") on node \"crc\" DevicePath \"\"" Mar 12 00:11:52 crc kubenswrapper[4948]: I0312 00:11:52.408873 4948 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/f90934d6-9305-4513-8d84-cd6a6fcc4a59-v4-0-config-system-ocp-branding-template\") on node \"crc\" DevicePath \"\"" Mar 12 00:11:52 crc kubenswrapper[4948]: I0312 00:11:52.408892 4948 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/f90934d6-9305-4513-8d84-cd6a6fcc4a59-v4-0-config-user-template-error\") on node \"crc\" DevicePath \"\"" Mar 12 00:11:52 crc kubenswrapper[4948]: I0312 00:11:52.408911 4948 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/f90934d6-9305-4513-8d84-cd6a6fcc4a59-v4-0-config-system-router-certs\") on node \"crc\" DevicePath \"\"" Mar 12 00:11:52 crc kubenswrapper[4948]: I0312 00:11:52.408929 4948 reconciler_common.go:293] "Volume detached for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/f90934d6-9305-4513-8d84-cd6a6fcc4a59-v4-0-config-system-service-ca\") on node \"crc\" DevicePath \"\"" Mar 12 00:11:52 crc kubenswrapper[4948]: I0312 00:11:52.408948 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dgm4m\" (UniqueName: \"kubernetes.io/projected/f90934d6-9305-4513-8d84-cd6a6fcc4a59-kube-api-access-dgm4m\") on node \"crc\" DevicePath \"\"" Mar 12 00:11:52 crc kubenswrapper[4948]: I0312 00:11:52.514540 4948 generic.go:334] "Generic (PLEG): container finished" podID="f90934d6-9305-4513-8d84-cd6a6fcc4a59" containerID="56f73e2a84df78402185d6ddeb2c483aaed0eb93cd03771074df5b28ef703b8a" exitCode=0 Mar 12 00:11:52 crc kubenswrapper[4948]: I0312 00:11:52.514588 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-qsf8x" event={"ID":"f90934d6-9305-4513-8d84-cd6a6fcc4a59","Type":"ContainerDied","Data":"56f73e2a84df78402185d6ddeb2c483aaed0eb93cd03771074df5b28ef703b8a"} Mar 12 00:11:52 crc kubenswrapper[4948]: I0312 00:11:52.514620 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-558db77b4-qsf8x" event={"ID":"f90934d6-9305-4513-8d84-cd6a6fcc4a59","Type":"ContainerDied","Data":"c684b0e15925fe0fa95fd90d3a1677094513bf25ca14305b5365b19e206e8576"} Mar 12 00:11:52 crc kubenswrapper[4948]: I0312 00:11:52.514640 4948 scope.go:117] "RemoveContainer" containerID="56f73e2a84df78402185d6ddeb2c483aaed0eb93cd03771074df5b28ef703b8a" Mar 12 00:11:52 crc kubenswrapper[4948]: I0312 00:11:52.514642 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-558db77b4-qsf8x" Mar 12 00:11:52 crc kubenswrapper[4948]: I0312 00:11:52.515262 4948 status_manager.go:851] "Failed to get status for pod" podUID="f90934d6-9305-4513-8d84-cd6a6fcc4a59" pod="openshift-authentication/oauth-openshift-558db77b4-qsf8x" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-qsf8x\": dial tcp 38.102.83.180:6443: connect: connection refused" Mar 12 00:11:52 crc kubenswrapper[4948]: I0312 00:11:52.515558 4948 status_manager.go:851] "Failed to get status for pod" podUID="62c7736b-4d37-43bf-8ee8-f1e5c3ac3a37" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Mar 12 00:11:52 crc kubenswrapper[4948]: I0312 00:11:52.547949 4948 scope.go:117] "RemoveContainer" containerID="56f73e2a84df78402185d6ddeb2c483aaed0eb93cd03771074df5b28ef703b8a" Mar 12 00:11:52 crc kubenswrapper[4948]: E0312 00:11:52.548658 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"56f73e2a84df78402185d6ddeb2c483aaed0eb93cd03771074df5b28ef703b8a\": container with ID starting with 56f73e2a84df78402185d6ddeb2c483aaed0eb93cd03771074df5b28ef703b8a not found: ID does not exist" containerID="56f73e2a84df78402185d6ddeb2c483aaed0eb93cd03771074df5b28ef703b8a" Mar 12 00:11:52 crc kubenswrapper[4948]: I0312 00:11:52.548746 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"56f73e2a84df78402185d6ddeb2c483aaed0eb93cd03771074df5b28ef703b8a"} err="failed to get container status \"56f73e2a84df78402185d6ddeb2c483aaed0eb93cd03771074df5b28ef703b8a\": rpc error: code = NotFound desc = could not find container \"56f73e2a84df78402185d6ddeb2c483aaed0eb93cd03771074df5b28ef703b8a\": container with ID starting with 56f73e2a84df78402185d6ddeb2c483aaed0eb93cd03771074df5b28ef703b8a not found: ID does not exist" Mar 12 00:11:52 crc kubenswrapper[4948]: I0312 00:11:52.551219 4948 status_manager.go:851] "Failed to get status for pod" podUID="f90934d6-9305-4513-8d84-cd6a6fcc4a59" pod="openshift-authentication/oauth-openshift-558db77b4-qsf8x" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-qsf8x\": dial tcp 38.102.83.180:6443: connect: connection refused" Mar 12 00:11:52 crc kubenswrapper[4948]: I0312 00:11:52.552482 4948 status_manager.go:851] "Failed to get status for pod" podUID="62c7736b-4d37-43bf-8ee8-f1e5c3ac3a37" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Mar 12 00:11:54 crc kubenswrapper[4948]: E0312 00:11:54.092244 4948 event.go:368] "Unable to write event (may retry after sleeping)" err="Post \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/events\": dial tcp 38.102.83.180:6443: connect: connection refused" event="&Event{ObjectMeta:{kube-apiserver-startup-monitor-crc.189bef964aea31e5 openshift-kube-apiserver 0 0001-01-01 00:00:00 +0000 UTC map[] map[] [] [] []},InvolvedObject:ObjectReference{Kind:Pod,Namespace:openshift-kube-apiserver,Name:kube-apiserver-startup-monitor-crc,UID:f85e55b1a89d02b0cb034b1ea31ed45a,APIVersion:v1,ResourceVersion:,FieldPath:spec.containers{startup-monitor},},Reason:Pulled,Message:Container image \"quay.io/crcont/openshift-crc-cluster-kube-apiserver-operator@sha256:9f36dc276e27753fc478274c7f7814a4f8945c987117ee1ea3b8e6355e6d7462\" already present on machine,Source:EventSource{Component:kubelet,Host:crc,},FirstTimestamp:2026-03-12 00:11:47.081417189 +0000 UTC m=+306.537020937,LastTimestamp:2026-03-12 00:11:47.081417189 +0000 UTC m=+306.537020937,Count:1,Type:Normal,EventTime:0001-01-01 00:00:00 +0000 UTC,Series:nil,Action:,Related:nil,ReportingController:kubelet,ReportingInstance:crc,}" Mar 12 00:11:56 crc kubenswrapper[4948]: E0312 00:11:56.359687 4948 desired_state_of_world_populator.go:312] "Error processing volume" err="error processing PVC openshift-image-registry/crc-image-registry-storage: failed to fetch PVC from API server: Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-image-registry/persistentvolumeclaims/crc-image-registry-storage\": dial tcp 38.102.83.180:6443: connect: connection refused" pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" volumeName="registry-storage" Mar 12 00:11:56 crc kubenswrapper[4948]: E0312 00:11:56.591824 4948 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.180:6443: connect: connection refused" Mar 12 00:11:56 crc kubenswrapper[4948]: E0312 00:11:56.592517 4948 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.180:6443: connect: connection refused" Mar 12 00:11:56 crc kubenswrapper[4948]: E0312 00:11:56.593243 4948 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.180:6443: connect: connection refused" Mar 12 00:11:56 crc kubenswrapper[4948]: E0312 00:11:56.594001 4948 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.180:6443: connect: connection refused" Mar 12 00:11:56 crc kubenswrapper[4948]: E0312 00:11:56.594700 4948 controller.go:195] "Failed to update lease" err="Put \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.180:6443: connect: connection refused" Mar 12 00:11:56 crc kubenswrapper[4948]: I0312 00:11:56.594766 4948 controller.go:115] "failed to update lease using latest lease, fallback to ensure lease" err="failed 5 attempts to update lease" Mar 12 00:11:56 crc kubenswrapper[4948]: E0312 00:11:56.595206 4948 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.180:6443: connect: connection refused" interval="200ms" Mar 12 00:11:56 crc kubenswrapper[4948]: E0312 00:11:56.796588 4948 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.180:6443: connect: connection refused" interval="400ms" Mar 12 00:11:57 crc kubenswrapper[4948]: E0312 00:11:57.197580 4948 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.180:6443: connect: connection refused" interval="800ms" Mar 12 00:11:57 crc kubenswrapper[4948]: E0312 00:11:57.998754 4948 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.180:6443: connect: connection refused" interval="1.6s" Mar 12 00:11:58 crc kubenswrapper[4948]: I0312 00:11:58.312368 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 00:11:58 crc kubenswrapper[4948]: I0312 00:11:58.313361 4948 status_manager.go:851] "Failed to get status for pod" podUID="f90934d6-9305-4513-8d84-cd6a6fcc4a59" pod="openshift-authentication/oauth-openshift-558db77b4-qsf8x" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-qsf8x\": dial tcp 38.102.83.180:6443: connect: connection refused" Mar 12 00:11:58 crc kubenswrapper[4948]: I0312 00:11:58.313942 4948 status_manager.go:851] "Failed to get status for pod" podUID="62c7736b-4d37-43bf-8ee8-f1e5c3ac3a37" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Mar 12 00:11:58 crc kubenswrapper[4948]: I0312 00:11:58.328813 4948 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d2efff53-8e12-41a1-bcb1-9a1d08c42a4b" Mar 12 00:11:58 crc kubenswrapper[4948]: I0312 00:11:58.328868 4948 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d2efff53-8e12-41a1-bcb1-9a1d08c42a4b" Mar 12 00:11:58 crc kubenswrapper[4948]: E0312 00:11:58.329879 4948 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 00:11:58 crc kubenswrapper[4948]: I0312 00:11:58.330671 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 00:11:58 crc kubenswrapper[4948]: I0312 00:11:58.557772 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"da6760c7cd985e8c6a9d3b3fb1bcb81c6d94746542b2bbaa9bf2702ba1ff6adc"} Mar 12 00:11:59 crc kubenswrapper[4948]: I0312 00:11:59.565343 4948 generic.go:334] "Generic (PLEG): container finished" podID="71bb4a3aecc4ba5b26c4b7318770ce13" containerID="19189b20f28d14258bd5d077ee6791eb3d93023d1d944c66969682eafbacf7ce" exitCode=0 Mar 12 00:11:59 crc kubenswrapper[4948]: I0312 00:11:59.565395 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerDied","Data":"19189b20f28d14258bd5d077ee6791eb3d93023d1d944c66969682eafbacf7ce"} Mar 12 00:11:59 crc kubenswrapper[4948]: I0312 00:11:59.566544 4948 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d2efff53-8e12-41a1-bcb1-9a1d08c42a4b" Mar 12 00:11:59 crc kubenswrapper[4948]: I0312 00:11:59.566556 4948 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d2efff53-8e12-41a1-bcb1-9a1d08c42a4b" Mar 12 00:11:59 crc kubenswrapper[4948]: E0312 00:11:59.567023 4948 mirror_client.go:138] "Failed deleting a mirror pod" err="Delete \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/kube-apiserver-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 00:11:59 crc kubenswrapper[4948]: I0312 00:11:59.567198 4948 status_manager.go:851] "Failed to get status for pod" podUID="f90934d6-9305-4513-8d84-cd6a6fcc4a59" pod="openshift-authentication/oauth-openshift-558db77b4-qsf8x" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-authentication/pods/oauth-openshift-558db77b4-qsf8x\": dial tcp 38.102.83.180:6443: connect: connection refused" Mar 12 00:11:59 crc kubenswrapper[4948]: I0312 00:11:59.567739 4948 status_manager.go:851] "Failed to get status for pod" podUID="62c7736b-4d37-43bf-8ee8-f1e5c3ac3a37" pod="openshift-kube-apiserver/installer-9-crc" err="Get \"https://api-int.crc.testing:6443/api/v1/namespaces/openshift-kube-apiserver/pods/installer-9-crc\": dial tcp 38.102.83.180:6443: connect: connection refused" Mar 12 00:11:59 crc kubenswrapper[4948]: E0312 00:11:59.600272 4948 controller.go:145] "Failed to ensure lease exists, will retry" err="Get \"https://api-int.crc.testing:6443/apis/coordination.k8s.io/v1/namespaces/kube-node-lease/leases/crc?timeout=10s\": dial tcp 38.102.83.180:6443: connect: connection refused" interval="3.2s" Mar 12 00:12:00 crc kubenswrapper[4948]: I0312 00:12:00.578971 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"df0ae8db5f73bbc97da61be18ab7bdc40d4bffc016fce300b85546ddc42ba3a3"} Mar 12 00:12:00 crc kubenswrapper[4948]: I0312 00:12:00.579012 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"d70334417e3d5ad6f2bfd836c8841578f8edd2347aecc4356d9b47fdc916f44c"} Mar 12 00:12:00 crc kubenswrapper[4948]: I0312 00:12:00.579021 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"ac1f9c2bad1abc718f780933e916d79f7f9d2f7e5d1b013f6e874b30ad43226d"} Mar 12 00:12:01 crc kubenswrapper[4948]: I0312 00:12:01.595225 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"97f22e96d8986eb7b4c45969fae4131d4058a7602a448fb1cf72fb6fcd4b53cb"} Mar 12 00:12:01 crc kubenswrapper[4948]: I0312 00:12:01.595780 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-apiserver/kube-apiserver-crc" event={"ID":"71bb4a3aecc4ba5b26c4b7318770ce13","Type":"ContainerStarted","Data":"704724493a76981dfe3ff44bd35e89da5b75be56a2bcee3ddf6f5c6497f5c50f"} Mar 12 00:12:01 crc kubenswrapper[4948]: I0312 00:12:01.595497 4948 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d2efff53-8e12-41a1-bcb1-9a1d08c42a4b" Mar 12 00:12:01 crc kubenswrapper[4948]: I0312 00:12:01.595845 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 00:12:01 crc kubenswrapper[4948]: I0312 00:12:01.595852 4948 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d2efff53-8e12-41a1-bcb1-9a1d08c42a4b" Mar 12 00:12:01 crc kubenswrapper[4948]: I0312 00:12:01.598528 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 12 00:12:01 crc kubenswrapper[4948]: I0312 00:12:01.599049 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Mar 12 00:12:01 crc kubenswrapper[4948]: I0312 00:12:01.599094 4948 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="fb45482a404d6696cfbc3919997e53146b7b9dd1536056a55374574ae01ff193" exitCode=1 Mar 12 00:12:01 crc kubenswrapper[4948]: I0312 00:12:01.599121 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"fb45482a404d6696cfbc3919997e53146b7b9dd1536056a55374574ae01ff193"} Mar 12 00:12:01 crc kubenswrapper[4948]: I0312 00:12:01.599606 4948 scope.go:117] "RemoveContainer" containerID="fb45482a404d6696cfbc3919997e53146b7b9dd1536056a55374574ae01ff193" Mar 12 00:12:01 crc kubenswrapper[4948]: I0312 00:12:01.878516 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 12 00:12:02 crc kubenswrapper[4948]: I0312 00:12:02.608907 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 12 00:12:02 crc kubenswrapper[4948]: I0312 00:12:02.609567 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Mar 12 00:12:02 crc kubenswrapper[4948]: I0312 00:12:02.609619 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"6d4ce0622c9e3a674aaaf793d0d56390132a140e51a9c7f4d37a94db9dce816f"} Mar 12 00:12:03 crc kubenswrapper[4948]: I0312 00:12:03.221220 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 12 00:12:03 crc kubenswrapper[4948]: I0312 00:12:03.221412 4948 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Mar 12 00:12:03 crc kubenswrapper[4948]: I0312 00:12:03.221762 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Mar 12 00:12:03 crc kubenswrapper[4948]: I0312 00:12:03.331517 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 00:12:03 crc kubenswrapper[4948]: I0312 00:12:03.331557 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 00:12:03 crc kubenswrapper[4948]: I0312 00:12:03.336041 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 00:12:06 crc kubenswrapper[4948]: I0312 00:12:06.608683 4948 kubelet.go:1914] "Deleted mirror pod because it is outdated" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 00:12:06 crc kubenswrapper[4948]: I0312 00:12:06.639052 4948 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d2efff53-8e12-41a1-bcb1-9a1d08c42a4b" Mar 12 00:12:06 crc kubenswrapper[4948]: I0312 00:12:06.639115 4948 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d2efff53-8e12-41a1-bcb1-9a1d08c42a4b" Mar 12 00:12:06 crc kubenswrapper[4948]: I0312 00:12:06.643583 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 00:12:06 crc kubenswrapper[4948]: I0312 00:12:06.646122 4948 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="28681f8a-fd34-4f03-a522-44fcea266158" Mar 12 00:12:07 crc kubenswrapper[4948]: I0312 00:12:07.644272 4948 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d2efff53-8e12-41a1-bcb1-9a1d08c42a4b" Mar 12 00:12:07 crc kubenswrapper[4948]: I0312 00:12:07.644603 4948 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d2efff53-8e12-41a1-bcb1-9a1d08c42a4b" Mar 12 00:12:11 crc kubenswrapper[4948]: I0312 00:12:11.342870 4948 status_manager.go:861] "Pod was deleted and then recreated, skipping status update" pod="openshift-kube-apiserver/kube-apiserver-crc" oldPodUID="71bb4a3aecc4ba5b26c4b7318770ce13" podUID="28681f8a-fd34-4f03-a522-44fcea266158" Mar 12 00:12:11 crc kubenswrapper[4948]: I0312 00:12:11.878625 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 12 00:12:13 crc kubenswrapper[4948]: I0312 00:12:13.221537 4948 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Mar 12 00:12:13 crc kubenswrapper[4948]: I0312 00:12:13.221659 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Mar 12 00:12:15 crc kubenswrapper[4948]: I0312 00:12:15.617158 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 12 00:12:16 crc kubenswrapper[4948]: I0312 00:12:16.546771 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"multus-daemon-config" Mar 12 00:12:16 crc kubenswrapper[4948]: I0312 00:12:16.903524 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"trusted-ca-bundle" Mar 12 00:12:16 crc kubenswrapper[4948]: I0312 00:12:16.977112 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-metrics-certs-default" Mar 12 00:12:17 crc kubenswrapper[4948]: I0312 00:12:17.151447 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"kube-root-ca.crt" Mar 12 00:12:17 crc kubenswrapper[4948]: I0312 00:12:17.305394 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"env-overrides" Mar 12 00:12:17 crc kubenswrapper[4948]: I0312 00:12:17.482717 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"pprof-cert" Mar 12 00:12:17 crc kubenswrapper[4948]: I0312 00:12:17.520362 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"cluster-samples-operator-dockercfg-xpp9w" Mar 12 00:12:17 crc kubenswrapper[4948]: I0312 00:12:17.755680 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-operator-config" Mar 12 00:12:18 crc kubenswrapper[4948]: I0312 00:12:18.148146 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"kube-root-ca.crt" Mar 12 00:12:18 crc kubenswrapper[4948]: I0312 00:12:18.185064 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-stats-default" Mar 12 00:12:18 crc kubenswrapper[4948]: I0312 00:12:18.905117 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-default-metrics-tls" Mar 12 00:12:19 crc kubenswrapper[4948]: I0312 00:12:19.044743 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"metrics-tls" Mar 12 00:12:19 crc kubenswrapper[4948]: I0312 00:12:19.080008 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"node-bootstrapper-token" Mar 12 00:12:19 crc kubenswrapper[4948]: I0312 00:12:19.186602 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"openshift-controller-manager-sa-dockercfg-msq4c" Mar 12 00:12:19 crc kubenswrapper[4948]: I0312 00:12:19.201016 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"encryption-config-1" Mar 12 00:12:19 crc kubenswrapper[4948]: I0312 00:12:19.384055 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"service-ca-dockercfg-pn86c" Mar 12 00:12:19 crc kubenswrapper[4948]: I0312 00:12:19.699527 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-tls" Mar 12 00:12:19 crc kubenswrapper[4948]: I0312 00:12:19.712846 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"openshift-config-operator-dockercfg-7pc5z" Mar 12 00:12:19 crc kubenswrapper[4948]: I0312 00:12:19.748622 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"packageserver-service-cert" Mar 12 00:12:20 crc kubenswrapper[4948]: I0312 00:12:20.062813 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"oauth-apiserver-sa-dockercfg-6r2bq" Mar 12 00:12:20 crc kubenswrapper[4948]: I0312 00:12:20.100198 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-node-identity"/"network-node-identity-cert" Mar 12 00:12:20 crc kubenswrapper[4948]: I0312 00:12:20.274892 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-tls" Mar 12 00:12:20 crc kubenswrapper[4948]: I0312 00:12:20.430699 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-dockercfg-x57mr" Mar 12 00:12:20 crc kubenswrapper[4948]: I0312 00:12:20.777528 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns-operator"/"openshift-service-ca.crt" Mar 12 00:12:20 crc kubenswrapper[4948]: I0312 00:12:20.822575 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"trusted-ca" Mar 12 00:12:20 crc kubenswrapper[4948]: I0312 00:12:20.843629 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mco-proxy-tls" Mar 12 00:12:20 crc kubenswrapper[4948]: I0312 00:12:20.864634 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"default-cni-sysctl-allowlist" Mar 12 00:12:20 crc kubenswrapper[4948]: I0312 00:12:20.950066 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-serving-cert" Mar 12 00:12:21 crc kubenswrapper[4948]: I0312 00:12:21.022031 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"oauth-serving-cert" Mar 12 00:12:21 crc kubenswrapper[4948]: I0312 00:12:21.220505 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-serving-cert" Mar 12 00:12:21 crc kubenswrapper[4948]: I0312 00:12:21.432628 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-root-ca.crt" Mar 12 00:12:21 crc kubenswrapper[4948]: I0312 00:12:21.503028 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"machine-api-operator-images" Mar 12 00:12:21 crc kubenswrapper[4948]: I0312 00:12:21.605963 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"trusted-ca" Mar 12 00:12:21 crc kubenswrapper[4948]: I0312 00:12:21.668443 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"encryption-config-1" Mar 12 00:12:21 crc kubenswrapper[4948]: I0312 00:12:21.670182 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"machine-approver-config" Mar 12 00:12:21 crc kubenswrapper[4948]: I0312 00:12:21.701469 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"etcd-client" Mar 12 00:12:21 crc kubenswrapper[4948]: I0312 00:12:21.867420 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"openshift-service-ca.crt" Mar 12 00:12:21 crc kubenswrapper[4948]: I0312 00:12:21.884147 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"audit-1" Mar 12 00:12:21 crc kubenswrapper[4948]: I0312 00:12:21.974710 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"kube-root-ca.crt" Mar 12 00:12:22 crc kubenswrapper[4948]: I0312 00:12:22.029678 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"kube-root-ca.crt" Mar 12 00:12:22 crc kubenswrapper[4948]: I0312 00:12:22.032478 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-config-operator"/"config-operator-serving-cert" Mar 12 00:12:22 crc kubenswrapper[4948]: I0312 00:12:22.060254 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-config" Mar 12 00:12:22 crc kubenswrapper[4948]: I0312 00:12:22.112613 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-tls" Mar 12 00:12:22 crc kubenswrapper[4948]: I0312 00:12:22.180611 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"openshift-service-ca.crt" Mar 12 00:12:22 crc kubenswrapper[4948]: I0312 00:12:22.267409 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"route-controller-manager-sa-dockercfg-h2zr2" Mar 12 00:12:22 crc kubenswrapper[4948]: I0312 00:12:22.304885 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"serving-cert" Mar 12 00:12:22 crc kubenswrapper[4948]: I0312 00:12:22.374499 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-root-ca.crt" Mar 12 00:12:22 crc kubenswrapper[4948]: I0312 00:12:22.387639 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"dns-default" Mar 12 00:12:22 crc kubenswrapper[4948]: I0312 00:12:22.438714 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"config" Mar 12 00:12:22 crc kubenswrapper[4948]: I0312 00:12:22.523478 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"trusted-ca" Mar 12 00:12:22 crc kubenswrapper[4948]: I0312 00:12:22.575782 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-config" Mar 12 00:12:22 crc kubenswrapper[4948]: I0312 00:12:22.576193 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"iptables-alerter-script" Mar 12 00:12:22 crc kubenswrapper[4948]: I0312 00:12:22.617994 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-config" Mar 12 00:12:22 crc kubenswrapper[4948]: I0312 00:12:22.659442 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager"/"serving-cert" Mar 12 00:12:22 crc kubenswrapper[4948]: I0312 00:12:22.696606 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"openshift-service-ca.crt" Mar 12 00:12:22 crc kubenswrapper[4948]: I0312 00:12:22.769126 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-console"/"networking-console-plugin-cert" Mar 12 00:12:22 crc kubenswrapper[4948]: I0312 00:12:22.828595 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-tls" Mar 12 00:12:22 crc kubenswrapper[4948]: I0312 00:12:22.882577 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"cluster-version-operator-serving-cert" Mar 12 00:12:22 crc kubenswrapper[4948]: I0312 00:12:22.911032 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"client-ca" Mar 12 00:12:22 crc kubenswrapper[4948]: I0312 00:12:22.928871 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-dockercfg-f62pw" Mar 12 00:12:23 crc kubenswrapper[4948]: I0312 00:12:23.140101 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"serving-cert" Mar 12 00:12:23 crc kubenswrapper[4948]: I0312 00:12:23.176135 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"openshift-apiserver-sa-dockercfg-djjff" Mar 12 00:12:23 crc kubenswrapper[4948]: I0312 00:12:23.218100 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-root-ca.crt" Mar 12 00:12:23 crc kubenswrapper[4948]: I0312 00:12:23.221407 4948 patch_prober.go:28] interesting pod/kube-controller-manager-crc container/kube-controller-manager namespace/openshift-kube-controller-manager: Startup probe status=failure output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" start-of-body= Mar 12 00:12:23 crc kubenswrapper[4948]: I0312 00:12:23.221460 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" probeResult="failure" output="Get \"https://192.168.126.11:10257/healthz\": dial tcp 192.168.126.11:10257: connect: connection refused" Mar 12 00:12:23 crc kubenswrapper[4948]: I0312 00:12:23.221507 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 12 00:12:23 crc kubenswrapper[4948]: I0312 00:12:23.222103 4948 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="kube-controller-manager" containerStatusID={"Type":"cri-o","ID":"6d4ce0622c9e3a674aaaf793d0d56390132a140e51a9c7f4d37a94db9dce816f"} pod="openshift-kube-controller-manager/kube-controller-manager-crc" containerMessage="Container kube-controller-manager failed startup probe, will be restarted" Mar 12 00:12:23 crc kubenswrapper[4948]: I0312 00:12:23.222198 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-controller-manager/kube-controller-manager-crc" podUID="f614b9022728cf315e60c057852e563e" containerName="kube-controller-manager" containerID="cri-o://6d4ce0622c9e3a674aaaf793d0d56390132a140e51a9c7f4d37a94db9dce816f" gracePeriod=30 Mar 12 00:12:23 crc kubenswrapper[4948]: I0312 00:12:23.235144 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"registry-dockercfg-kzzsd" Mar 12 00:12:23 crc kubenswrapper[4948]: I0312 00:12:23.235397 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"machine-config-operator-images" Mar 12 00:12:23 crc kubenswrapper[4948]: I0312 00:12:23.268214 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-config" Mar 12 00:12:23 crc kubenswrapper[4948]: I0312 00:12:23.309036 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"kube-root-ca.crt" Mar 12 00:12:23 crc kubenswrapper[4948]: I0312 00:12:23.315077 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"openshift-service-ca.crt" Mar 12 00:12:23 crc kubenswrapper[4948]: I0312 00:12:23.413650 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-secret" Mar 12 00:12:23 crc kubenswrapper[4948]: I0312 00:12:23.425440 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-diagnostics"/"kube-root-ca.crt" Mar 12 00:12:23 crc kubenswrapper[4948]: I0312 00:12:23.482952 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-controller-manager-operator"/"kube-controller-manager-operator-dockercfg-gkqpw" Mar 12 00:12:23 crc kubenswrapper[4948]: I0312 00:12:23.489507 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"openshift-service-ca.crt" Mar 12 00:12:23 crc kubenswrapper[4948]: I0312 00:12:23.521758 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-version"/"default-dockercfg-gxtc4" Mar 12 00:12:23 crc kubenswrapper[4948]: I0312 00:12:23.539166 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-serving-cert" Mar 12 00:12:23 crc kubenswrapper[4948]: I0312 00:12:23.550054 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"kube-root-ca.crt" Mar 12 00:12:23 crc kubenswrapper[4948]: I0312 00:12:23.682167 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"trusted-ca-bundle" Mar 12 00:12:23 crc kubenswrapper[4948]: I0312 00:12:23.718847 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"openshift-service-ca.crt" Mar 12 00:12:23 crc kubenswrapper[4948]: I0312 00:12:23.768473 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"image-registry-certificates" Mar 12 00:12:23 crc kubenswrapper[4948]: I0312 00:12:23.769878 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-service-ca.crt" Mar 12 00:12:23 crc kubenswrapper[4948]: I0312 00:12:23.775295 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"authentication-operator-config" Mar 12 00:12:23 crc kubenswrapper[4948]: I0312 00:12:23.780430 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ac-dockercfg-9lkdf" Mar 12 00:12:23 crc kubenswrapper[4948]: I0312 00:12:23.837616 4948 reflector.go:368] Caches populated for *v1.Secret from object-"hostpath-provisioner"/"csi-hostpath-provisioner-sa-dockercfg-qd74k" Mar 12 00:12:23 crc kubenswrapper[4948]: I0312 00:12:23.837724 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"image-import-ca" Mar 12 00:12:23 crc kubenswrapper[4948]: I0312 00:12:23.947513 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager-operator"/"kube-root-ca.crt" Mar 12 00:12:24 crc kubenswrapper[4948]: I0312 00:12:24.104672 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"kube-root-ca.crt" Mar 12 00:12:24 crc kubenswrapper[4948]: I0312 00:12:24.104814 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"etcd-serving-ca" Mar 12 00:12:24 crc kubenswrapper[4948]: I0312 00:12:24.205745 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"openshift-service-ca.crt" Mar 12 00:12:24 crc kubenswrapper[4948]: I0312 00:12:24.314287 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"kube-root-ca.crt" Mar 12 00:12:24 crc kubenswrapper[4948]: I0312 00:12:24.361250 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"control-plane-machine-set-operator-dockercfg-k9rxt" Mar 12 00:12:24 crc kubenswrapper[4948]: I0312 00:12:24.451862 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-machine-approver"/"machine-approver-sa-dockercfg-nl2j4" Mar 12 00:12:24 crc kubenswrapper[4948]: I0312 00:12:24.485265 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serviceaccount-dockercfg-rq7zk" Mar 12 00:12:24 crc kubenswrapper[4948]: I0312 00:12:24.544727 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"kube-root-ca.crt" Mar 12 00:12:24 crc kubenswrapper[4948]: I0312 00:12:24.641975 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"openshift-service-ca.crt" Mar 12 00:12:24 crc kubenswrapper[4948]: I0312 00:12:24.696633 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"installation-pull-secrets" Mar 12 00:12:24 crc kubenswrapper[4948]: I0312 00:12:24.744231 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"kube-root-ca.crt" Mar 12 00:12:24 crc kubenswrapper[4948]: I0312 00:12:24.830349 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-serving-cert" Mar 12 00:12:24 crc kubenswrapper[4948]: I0312 00:12:24.833875 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-controller-manager-operator"/"openshift-controller-manager-operator-dockercfg-vw8fw" Mar 12 00:12:24 crc kubenswrapper[4948]: I0312 00:12:24.847500 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"trusted-ca-bundle" Mar 12 00:12:24 crc kubenswrapper[4948]: I0312 00:12:24.982758 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 12 00:12:24 crc kubenswrapper[4948]: I0312 00:12:24.995408 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"signing-cabundle" Mar 12 00:12:25 crc kubenswrapper[4948]: I0312 00:12:25.001015 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"kube-root-ca.crt" Mar 12 00:12:25 crc kubenswrapper[4948]: I0312 00:12:25.020986 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-apiserver-operator"/"kube-apiserver-operator-serving-cert" Mar 12 00:12:25 crc kubenswrapper[4948]: I0312 00:12:25.022045 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"openshift-service-ca.crt" Mar 12 00:12:25 crc kubenswrapper[4948]: I0312 00:12:25.124098 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"node-resolver-dockercfg-kz9s7" Mar 12 00:12:25 crc kubenswrapper[4948]: I0312 00:12:25.157681 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"env-overrides" Mar 12 00:12:25 crc kubenswrapper[4948]: I0312 00:12:25.179738 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"metrics-daemon-sa-dockercfg-d427c" Mar 12 00:12:25 crc kubenswrapper[4948]: I0312 00:12:25.254779 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns"/"dns-dockercfg-jwfmh" Mar 12 00:12:25 crc kubenswrapper[4948]: I0312 00:12:25.292835 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"openshift-service-ca.crt" Mar 12 00:12:25 crc kubenswrapper[4948]: I0312 00:12:25.455952 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator"/"kube-root-ca.crt" Mar 12 00:12:25 crc kubenswrapper[4948]: I0312 00:12:25.457580 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"service-ca-operator-config" Mar 12 00:12:25 crc kubenswrapper[4948]: I0312 00:12:25.701590 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-daemon-dockercfg-r5tcq" Mar 12 00:12:25 crc kubenswrapper[4948]: I0312 00:12:25.792140 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"config" Mar 12 00:12:25 crc kubenswrapper[4948]: I0312 00:12:25.799062 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-control-plane-metrics-cert" Mar 12 00:12:25 crc kubenswrapper[4948]: I0312 00:12:25.811276 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"authentication-operator-dockercfg-mz9bj" Mar 12 00:12:25 crc kubenswrapper[4948]: I0312 00:12:25.831530 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-oauth-apiserver"/"serving-cert" Mar 12 00:12:25 crc kubenswrapper[4948]: I0312 00:12:25.954703 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"openshift-service-ca.crt" Mar 12 00:12:26 crc kubenswrapper[4948]: I0312 00:12:26.003807 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-config-operator"/"openshift-service-ca.crt" Mar 12 00:12:26 crc kubenswrapper[4948]: I0312 00:12:26.037725 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"openshift-service-ca.crt" Mar 12 00:12:26 crc kubenswrapper[4948]: I0312 00:12:26.044638 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-oauth-config" Mar 12 00:12:26 crc kubenswrapper[4948]: I0312 00:12:26.123116 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"openshift-service-ca.crt" Mar 12 00:12:26 crc kubenswrapper[4948]: I0312 00:12:26.307285 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"kube-scheduler-operator-serving-cert" Mar 12 00:12:26 crc kubenswrapper[4948]: I0312 00:12:26.317839 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"kube-root-ca.crt" Mar 12 00:12:26 crc kubenswrapper[4948]: I0312 00:12:26.363235 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"config" Mar 12 00:12:26 crc kubenswrapper[4948]: I0312 00:12:26.371513 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console-operator"/"console-operator-config" Mar 12 00:12:26 crc kubenswrapper[4948]: I0312 00:12:26.466002 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-cluster-samples-operator"/"samples-operator-tls" Mar 12 00:12:26 crc kubenswrapper[4948]: I0312 00:12:26.496661 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"proxy-tls" Mar 12 00:12:26 crc kubenswrapper[4948]: I0312 00:12:26.551891 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"client-ca" Mar 12 00:12:26 crc kubenswrapper[4948]: I0312 00:12:26.649157 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"openshift-service-ca.crt" Mar 12 00:12:26 crc kubenswrapper[4948]: I0312 00:12:26.668485 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-image-registry"/"kube-root-ca.crt" Mar 12 00:12:26 crc kubenswrapper[4948]: I0312 00:12:26.704931 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"ovnkube-identity-cm" Mar 12 00:12:26 crc kubenswrapper[4948]: I0312 00:12:26.723519 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"service-ca" Mar 12 00:12:26 crc kubenswrapper[4948]: I0312 00:12:26.790009 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"openshift-global-ca" Mar 12 00:12:26 crc kubenswrapper[4948]: I0312 00:12:26.828964 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"kube-root-ca.crt" Mar 12 00:12:26 crc kubenswrapper[4948]: I0312 00:12:26.863807 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-console"/"networking-console-plugin" Mar 12 00:12:26 crc kubenswrapper[4948]: I0312 00:12:26.901666 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication-operator"/"serving-cert" Mar 12 00:12:26 crc kubenswrapper[4948]: I0312 00:12:26.924761 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"default-dockercfg-2llfx" Mar 12 00:12:27 crc kubenswrapper[4948]: I0312 00:12:27.081003 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-admission-controller-secret" Mar 12 00:12:27 crc kubenswrapper[4948]: I0312 00:12:27.093197 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"serving-cert" Mar 12 00:12:27 crc kubenswrapper[4948]: I0312 00:12:27.153092 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"console-operator-dockercfg-4xjcr" Mar 12 00:12:27 crc kubenswrapper[4948]: I0312 00:12:27.191541 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-dockercfg-qt55r" Mar 12 00:12:27 crc kubenswrapper[4948]: I0312 00:12:27.209950 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-controller-dockercfg-c2lfx" Mar 12 00:12:27 crc kubenswrapper[4948]: I0312 00:12:27.210290 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-route-controller-manager"/"serving-cert" Mar 12 00:12:27 crc kubenswrapper[4948]: I0312 00:12:27.289132 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"package-server-manager-serving-cert" Mar 12 00:12:27 crc kubenswrapper[4948]: I0312 00:12:27.419722 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console-operator"/"serving-cert" Mar 12 00:12:27 crc kubenswrapper[4948]: I0312 00:12:27.452921 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-dns"/"kube-root-ca.crt" Mar 12 00:12:27 crc kubenswrapper[4948]: I0312 00:12:27.515130 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-dockercfg-zdk86" Mar 12 00:12:27 crc kubenswrapper[4948]: I0312 00:12:27.526275 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-dns-operator"/"dns-operator-dockercfg-9mqw5" Mar 12 00:12:27 crc kubenswrapper[4948]: I0312 00:12:27.621805 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-operator"/"kube-root-ca.crt" Mar 12 00:12:27 crc kubenswrapper[4948]: I0312 00:12:27.644768 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"multus-ancillary-tools-dockercfg-vnmsz" Mar 12 00:12:27 crc kubenswrapper[4948]: I0312 00:12:27.669379 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator-operator"/"kube-storage-version-migrator-operator-dockercfg-2bh8d" Mar 12 00:12:27 crc kubenswrapper[4948]: I0312 00:12:27.775231 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"kube-root-ca.crt" Mar 12 00:12:27 crc kubenswrapper[4948]: I0312 00:12:27.789670 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"openshift-service-ca.crt" Mar 12 00:12:27 crc kubenswrapper[4948]: I0312 00:12:27.795397 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"etcd-serving-ca" Mar 12 00:12:27 crc kubenswrapper[4948]: I0312 00:12:27.828966 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"kube-rbac-proxy" Mar 12 00:12:27 crc kubenswrapper[4948]: I0312 00:12:27.887394 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"kube-rbac-proxy" Mar 12 00:12:27 crc kubenswrapper[4948]: I0312 00:12:27.990391 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-tls" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.008792 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-network-operator"/"metrics-tls" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.010040 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-api"/"kube-rbac-proxy" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.010544 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"console-serving-cert" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.129343 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-multus"/"default-dockercfg-2q5b6" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.168405 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-client" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.191947 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"openshift-kube-scheduler-operator-config" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.281852 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"kube-root-ca.crt" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.309075 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-etcd-operator"/"etcd-operator-dockercfg-r9srn" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.318054 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-operator"/"kube-root-ca.crt" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.338258 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-operator-dockercfg-98p87" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.381490 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"kube-root-ca.crt" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.431116 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"image-registry-operator-tls" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.477585 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-samples-operator"/"openshift-service-ca.crt" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.514555 4948 reflector.go:368] Caches populated for *v1.Pod from pkg/kubelet/config/apiserver.go:66 Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.519381 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-kube-apiserver/kube-apiserver-crc","openshift-authentication/oauth-openshift-558db77b4-qsf8x"] Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.519447 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-authentication/oauth-openshift-7d9c768c99-hpghc","openshift-kube-apiserver/kube-apiserver-crc"] Mar 12 00:12:28 crc kubenswrapper[4948]: E0312 00:12:28.519628 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="62c7736b-4d37-43bf-8ee8-f1e5c3ac3a37" containerName="installer" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.519646 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="62c7736b-4d37-43bf-8ee8-f1e5c3ac3a37" containerName="installer" Mar 12 00:12:28 crc kubenswrapper[4948]: E0312 00:12:28.519689 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f90934d6-9305-4513-8d84-cd6a6fcc4a59" containerName="oauth-openshift" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.519699 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="f90934d6-9305-4513-8d84-cd6a6fcc4a59" containerName="oauth-openshift" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.519814 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="62c7736b-4d37-43bf-8ee8-f1e5c3ac3a37" containerName="installer" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.519833 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="f90934d6-9305-4513-8d84-cd6a6fcc4a59" containerName="oauth-openshift" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.519852 4948 kubelet.go:1909] "Trying to delete pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d2efff53-8e12-41a1-bcb1-9a1d08c42a4b" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.519875 4948 mirror_client.go:130] "Deleting a mirror pod" pod="openshift-kube-apiserver/kube-apiserver-crc" podUID="d2efff53-8e12-41a1-bcb1-9a1d08c42a4b" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.520241 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7d9c768c99-hpghc" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.522777 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-cliconfig" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.522963 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-session" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.523172 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-idp-0-file-data" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.524246 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-apiserver/kube-apiserver-crc" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.524912 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-router-certs" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.525056 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"openshift-service-ca.crt" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.525706 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"oauth-openshift-dockercfg-znhcc" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.526935 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"audit" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.527586 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-error" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.527694 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-serving-cert" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.527587 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-provider-selection" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.527914 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"kube-root-ca.crt" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.528403 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-service-ca" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.551454 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-user-template-login" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.552690 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication"/"v4-0-config-system-trusted-ca-bundle" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.570331 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-kube-apiserver/kube-apiserver-crc" podStartSLOduration=22.570295282 podStartE2EDuration="22.570295282s" podCreationTimestamp="2026-03-12 00:12:06 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:12:28.56337189 +0000 UTC m=+348.018975648" watchObservedRunningTime="2026-03-12 00:12:28.570295282 +0000 UTC m=+348.025899020" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.570770 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-authentication"/"v4-0-config-system-ocp-branding-template" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.582839 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"service-ca-bundle" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.597040 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ea491ee5-9d95-4a63-896b-4ebd3ceec50a-v4-0-config-system-router-certs\") pod \"oauth-openshift-7d9c768c99-hpghc\" (UID: \"ea491ee5-9d95-4a63-896b-4ebd3ceec50a\") " pod="openshift-authentication/oauth-openshift-7d9c768c99-hpghc" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.597249 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ea491ee5-9d95-4a63-896b-4ebd3ceec50a-v4-0-config-user-template-login\") pod \"oauth-openshift-7d9c768c99-hpghc\" (UID: \"ea491ee5-9d95-4a63-896b-4ebd3ceec50a\") " pod="openshift-authentication/oauth-openshift-7d9c768c99-hpghc" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.597410 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mkqj4\" (UniqueName: \"kubernetes.io/projected/ea491ee5-9d95-4a63-896b-4ebd3ceec50a-kube-api-access-mkqj4\") pod \"oauth-openshift-7d9c768c99-hpghc\" (UID: \"ea491ee5-9d95-4a63-896b-4ebd3ceec50a\") " pod="openshift-authentication/oauth-openshift-7d9c768c99-hpghc" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.597516 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ea491ee5-9d95-4a63-896b-4ebd3ceec50a-v4-0-config-system-session\") pod \"oauth-openshift-7d9c768c99-hpghc\" (UID: \"ea491ee5-9d95-4a63-896b-4ebd3ceec50a\") " pod="openshift-authentication/oauth-openshift-7d9c768c99-hpghc" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.597648 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ea491ee5-9d95-4a63-896b-4ebd3ceec50a-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7d9c768c99-hpghc\" (UID: \"ea491ee5-9d95-4a63-896b-4ebd3ceec50a\") " pod="openshift-authentication/oauth-openshift-7d9c768c99-hpghc" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.597711 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ea491ee5-9d95-4a63-896b-4ebd3ceec50a-v4-0-config-system-service-ca\") pod \"oauth-openshift-7d9c768c99-hpghc\" (UID: \"ea491ee5-9d95-4a63-896b-4ebd3ceec50a\") " pod="openshift-authentication/oauth-openshift-7d9c768c99-hpghc" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.597757 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ea491ee5-9d95-4a63-896b-4ebd3ceec50a-v4-0-config-user-template-error\") pod \"oauth-openshift-7d9c768c99-hpghc\" (UID: \"ea491ee5-9d95-4a63-896b-4ebd3ceec50a\") " pod="openshift-authentication/oauth-openshift-7d9c768c99-hpghc" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.597814 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ea491ee5-9d95-4a63-896b-4ebd3ceec50a-audit-policies\") pod \"oauth-openshift-7d9c768c99-hpghc\" (UID: \"ea491ee5-9d95-4a63-896b-4ebd3ceec50a\") " pod="openshift-authentication/oauth-openshift-7d9c768c99-hpghc" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.597897 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ea491ee5-9d95-4a63-896b-4ebd3ceec50a-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7d9c768c99-hpghc\" (UID: \"ea491ee5-9d95-4a63-896b-4ebd3ceec50a\") " pod="openshift-authentication/oauth-openshift-7d9c768c99-hpghc" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.598061 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ea491ee5-9d95-4a63-896b-4ebd3ceec50a-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7d9c768c99-hpghc\" (UID: \"ea491ee5-9d95-4a63-896b-4ebd3ceec50a\") " pod="openshift-authentication/oauth-openshift-7d9c768c99-hpghc" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.598131 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ea491ee5-9d95-4a63-896b-4ebd3ceec50a-audit-dir\") pod \"oauth-openshift-7d9c768c99-hpghc\" (UID: \"ea491ee5-9d95-4a63-896b-4ebd3ceec50a\") " pod="openshift-authentication/oauth-openshift-7d9c768c99-hpghc" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.598181 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ea491ee5-9d95-4a63-896b-4ebd3ceec50a-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7d9c768c99-hpghc\" (UID: \"ea491ee5-9d95-4a63-896b-4ebd3ceec50a\") " pod="openshift-authentication/oauth-openshift-7d9c768c99-hpghc" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.598273 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ea491ee5-9d95-4a63-896b-4ebd3ceec50a-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7d9c768c99-hpghc\" (UID: \"ea491ee5-9d95-4a63-896b-4ebd3ceec50a\") " pod="openshift-authentication/oauth-openshift-7d9c768c99-hpghc" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.598387 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ea491ee5-9d95-4a63-896b-4ebd3ceec50a-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7d9c768c99-hpghc\" (UID: \"ea491ee5-9d95-4a63-896b-4ebd3ceec50a\") " pod="openshift-authentication/oauth-openshift-7d9c768c99-hpghc" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.618575 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"node-ca-dockercfg-4777p" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.685969 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"mcc-proxy-tls" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.699180 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ea491ee5-9d95-4a63-896b-4ebd3ceec50a-audit-dir\") pod \"oauth-openshift-7d9c768c99-hpghc\" (UID: \"ea491ee5-9d95-4a63-896b-4ebd3ceec50a\") " pod="openshift-authentication/oauth-openshift-7d9c768c99-hpghc" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.699661 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ea491ee5-9d95-4a63-896b-4ebd3ceec50a-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7d9c768c99-hpghc\" (UID: \"ea491ee5-9d95-4a63-896b-4ebd3ceec50a\") " pod="openshift-authentication/oauth-openshift-7d9c768c99-hpghc" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.699452 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-dir\" (UniqueName: \"kubernetes.io/host-path/ea491ee5-9d95-4a63-896b-4ebd3ceec50a-audit-dir\") pod \"oauth-openshift-7d9c768c99-hpghc\" (UID: \"ea491ee5-9d95-4a63-896b-4ebd3ceec50a\") " pod="openshift-authentication/oauth-openshift-7d9c768c99-hpghc" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.700169 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ea491ee5-9d95-4a63-896b-4ebd3ceec50a-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7d9c768c99-hpghc\" (UID: \"ea491ee5-9d95-4a63-896b-4ebd3ceec50a\") " pod="openshift-authentication/oauth-openshift-7d9c768c99-hpghc" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.700438 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ea491ee5-9d95-4a63-896b-4ebd3ceec50a-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7d9c768c99-hpghc\" (UID: \"ea491ee5-9d95-4a63-896b-4ebd3ceec50a\") " pod="openshift-authentication/oauth-openshift-7d9c768c99-hpghc" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.700717 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-trusted-ca-bundle\" (UniqueName: \"kubernetes.io/configmap/ea491ee5-9d95-4a63-896b-4ebd3ceec50a-v4-0-config-system-trusted-ca-bundle\") pod \"oauth-openshift-7d9c768c99-hpghc\" (UID: \"ea491ee5-9d95-4a63-896b-4ebd3ceec50a\") " pod="openshift-authentication/oauth-openshift-7d9c768c99-hpghc" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.700719 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ea491ee5-9d95-4a63-896b-4ebd3ceec50a-v4-0-config-system-router-certs\") pod \"oauth-openshift-7d9c768c99-hpghc\" (UID: \"ea491ee5-9d95-4a63-896b-4ebd3ceec50a\") " pod="openshift-authentication/oauth-openshift-7d9c768c99-hpghc" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.700789 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ea491ee5-9d95-4a63-896b-4ebd3ceec50a-v4-0-config-user-template-login\") pod \"oauth-openshift-7d9c768c99-hpghc\" (UID: \"ea491ee5-9d95-4a63-896b-4ebd3ceec50a\") " pod="openshift-authentication/oauth-openshift-7d9c768c99-hpghc" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.700809 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mkqj4\" (UniqueName: \"kubernetes.io/projected/ea491ee5-9d95-4a63-896b-4ebd3ceec50a-kube-api-access-mkqj4\") pod \"oauth-openshift-7d9c768c99-hpghc\" (UID: \"ea491ee5-9d95-4a63-896b-4ebd3ceec50a\") " pod="openshift-authentication/oauth-openshift-7d9c768c99-hpghc" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.700852 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ea491ee5-9d95-4a63-896b-4ebd3ceec50a-v4-0-config-system-session\") pod \"oauth-openshift-7d9c768c99-hpghc\" (UID: \"ea491ee5-9d95-4a63-896b-4ebd3ceec50a\") " pod="openshift-authentication/oauth-openshift-7d9c768c99-hpghc" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.700897 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ea491ee5-9d95-4a63-896b-4ebd3ceec50a-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7d9c768c99-hpghc\" (UID: \"ea491ee5-9d95-4a63-896b-4ebd3ceec50a\") " pod="openshift-authentication/oauth-openshift-7d9c768c99-hpghc" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.700917 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ea491ee5-9d95-4a63-896b-4ebd3ceec50a-v4-0-config-system-service-ca\") pod \"oauth-openshift-7d9c768c99-hpghc\" (UID: \"ea491ee5-9d95-4a63-896b-4ebd3ceec50a\") " pod="openshift-authentication/oauth-openshift-7d9c768c99-hpghc" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.700932 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ea491ee5-9d95-4a63-896b-4ebd3ceec50a-v4-0-config-user-template-error\") pod \"oauth-openshift-7d9c768c99-hpghc\" (UID: \"ea491ee5-9d95-4a63-896b-4ebd3ceec50a\") " pod="openshift-authentication/oauth-openshift-7d9c768c99-hpghc" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.700954 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ea491ee5-9d95-4a63-896b-4ebd3ceec50a-audit-policies\") pod \"oauth-openshift-7d9c768c99-hpghc\" (UID: \"ea491ee5-9d95-4a63-896b-4ebd3ceec50a\") " pod="openshift-authentication/oauth-openshift-7d9c768c99-hpghc" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.700990 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ea491ee5-9d95-4a63-896b-4ebd3ceec50a-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7d9c768c99-hpghc\" (UID: \"ea491ee5-9d95-4a63-896b-4ebd3ceec50a\") " pod="openshift-authentication/oauth-openshift-7d9c768c99-hpghc" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.701056 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ea491ee5-9d95-4a63-896b-4ebd3ceec50a-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7d9c768c99-hpghc\" (UID: \"ea491ee5-9d95-4a63-896b-4ebd3ceec50a\") " pod="openshift-authentication/oauth-openshift-7d9c768c99-hpghc" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.701574 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-cliconfig\" (UniqueName: \"kubernetes.io/configmap/ea491ee5-9d95-4a63-896b-4ebd3ceec50a-v4-0-config-system-cliconfig\") pod \"oauth-openshift-7d9c768c99-hpghc\" (UID: \"ea491ee5-9d95-4a63-896b-4ebd3ceec50a\") " pod="openshift-authentication/oauth-openshift-7d9c768c99-hpghc" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.702020 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-service-ca\" (UniqueName: \"kubernetes.io/configmap/ea491ee5-9d95-4a63-896b-4ebd3ceec50a-v4-0-config-system-service-ca\") pod \"oauth-openshift-7d9c768c99-hpghc\" (UID: \"ea491ee5-9d95-4a63-896b-4ebd3ceec50a\") " pod="openshift-authentication/oauth-openshift-7d9c768c99-hpghc" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.702360 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca"/"signing-key" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.702415 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"audit-policies\" (UniqueName: \"kubernetes.io/configmap/ea491ee5-9d95-4a63-896b-4ebd3ceec50a-audit-policies\") pod \"oauth-openshift-7d9c768c99-hpghc\" (UID: \"ea491ee5-9d95-4a63-896b-4ebd3ceec50a\") " pod="openshift-authentication/oauth-openshift-7d9c768c99-hpghc" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.706654 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-error\" (UniqueName: \"kubernetes.io/secret/ea491ee5-9d95-4a63-896b-4ebd3ceec50a-v4-0-config-user-template-error\") pod \"oauth-openshift-7d9c768c99-hpghc\" (UID: \"ea491ee5-9d95-4a63-896b-4ebd3ceec50a\") " pod="openshift-authentication/oauth-openshift-7d9c768c99-hpghc" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.707223 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-idp-0-file-data\" (UniqueName: \"kubernetes.io/secret/ea491ee5-9d95-4a63-896b-4ebd3ceec50a-v4-0-config-user-idp-0-file-data\") pod \"oauth-openshift-7d9c768c99-hpghc\" (UID: \"ea491ee5-9d95-4a63-896b-4ebd3ceec50a\") " pod="openshift-authentication/oauth-openshift-7d9c768c99-hpghc" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.708071 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-provider-selection\" (UniqueName: \"kubernetes.io/secret/ea491ee5-9d95-4a63-896b-4ebd3ceec50a-v4-0-config-user-template-provider-selection\") pod \"oauth-openshift-7d9c768c99-hpghc\" (UID: \"ea491ee5-9d95-4a63-896b-4ebd3ceec50a\") " pod="openshift-authentication/oauth-openshift-7d9c768c99-hpghc" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.709123 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-router-certs\" (UniqueName: \"kubernetes.io/secret/ea491ee5-9d95-4a63-896b-4ebd3ceec50a-v4-0-config-system-router-certs\") pod \"oauth-openshift-7d9c768c99-hpghc\" (UID: \"ea491ee5-9d95-4a63-896b-4ebd3ceec50a\") " pod="openshift-authentication/oauth-openshift-7d9c768c99-hpghc" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.709942 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-session\" (UniqueName: \"kubernetes.io/secret/ea491ee5-9d95-4a63-896b-4ebd3ceec50a-v4-0-config-system-session\") pod \"oauth-openshift-7d9c768c99-hpghc\" (UID: \"ea491ee5-9d95-4a63-896b-4ebd3ceec50a\") " pod="openshift-authentication/oauth-openshift-7d9c768c99-hpghc" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.710256 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-serving-cert\" (UniqueName: \"kubernetes.io/secret/ea491ee5-9d95-4a63-896b-4ebd3ceec50a-v4-0-config-system-serving-cert\") pod \"oauth-openshift-7d9c768c99-hpghc\" (UID: \"ea491ee5-9d95-4a63-896b-4ebd3ceec50a\") " pod="openshift-authentication/oauth-openshift-7d9c768c99-hpghc" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.713880 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-system-ocp-branding-template\" (UniqueName: \"kubernetes.io/secret/ea491ee5-9d95-4a63-896b-4ebd3ceec50a-v4-0-config-system-ocp-branding-template\") pod \"oauth-openshift-7d9c768c99-hpghc\" (UID: \"ea491ee5-9d95-4a63-896b-4ebd3ceec50a\") " pod="openshift-authentication/oauth-openshift-7d9c768c99-hpghc" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.717456 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"v4-0-config-user-template-login\" (UniqueName: \"kubernetes.io/secret/ea491ee5-9d95-4a63-896b-4ebd3ceec50a-v4-0-config-user-template-login\") pod \"oauth-openshift-7d9c768c99-hpghc\" (UID: \"ea491ee5-9d95-4a63-896b-4ebd3ceec50a\") " pod="openshift-authentication/oauth-openshift-7d9c768c99-hpghc" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.731337 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mkqj4\" (UniqueName: \"kubernetes.io/projected/ea491ee5-9d95-4a63-896b-4ebd3ceec50a-kube-api-access-mkqj4\") pod \"oauth-openshift-7d9c768c99-hpghc\" (UID: \"ea491ee5-9d95-4a63-896b-4ebd3ceec50a\") " pod="openshift-authentication/oauth-openshift-7d9c768c99-hpghc" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.791418 4948 reflector.go:368] Caches populated for *v1.Node from k8s.io/client-go/informers/factory.go:160 Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.828427 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.842347 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-metrics" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.853406 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-authentication/oauth-openshift-7d9c768c99-hpghc" Mar 12 00:12:28 crc kubenswrapper[4948]: I0312 00:12:28.950806 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"cni-copy-resources" Mar 12 00:12:29 crc kubenswrapper[4948]: I0312 00:12:29.003963 4948 kubelet.go:2431] "SyncLoop REMOVE" source="file" pods=["openshift-kube-apiserver/kube-apiserver-startup-monitor-crc"] Mar 12 00:12:29 crc kubenswrapper[4948]: I0312 00:12:29.004272 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" containerID="cri-o://b8f42af8d5706c672759fe6b3ea65eaccb27bc3e80ed254e63660d83b69bfcde" gracePeriod=5 Mar 12 00:12:29 crc kubenswrapper[4948]: I0312 00:12:29.092954 4948 reflector.go:368] Caches populated for *v1.Service from k8s.io/client-go/informers/factory.go:160 Mar 12 00:12:29 crc kubenswrapper[4948]: I0312 00:12:29.104464 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-apiserver-operator"/"kube-root-ca.crt" Mar 12 00:12:29 crc kubenswrapper[4948]: I0312 00:12:29.300101 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-authentication/oauth-openshift-7d9c768c99-hpghc"] Mar 12 00:12:29 crc kubenswrapper[4948]: I0312 00:12:29.321638 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f90934d6-9305-4513-8d84-cd6a6fcc4a59" path="/var/lib/kubelet/pods/f90934d6-9305-4513-8d84-cd6a6fcc4a59/volumes" Mar 12 00:12:29 crc kubenswrapper[4948]: I0312 00:12:29.396181 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"marketplace-trusted-ca" Mar 12 00:12:29 crc kubenswrapper[4948]: I0312 00:12:29.503348 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"kube-root-ca.crt" Mar 12 00:12:29 crc kubenswrapper[4948]: I0312 00:12:29.730983 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-controller-manager-operator"/"kube-root-ca.crt" Mar 12 00:12:29 crc kubenswrapper[4948]: I0312 00:12:29.749816 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-control-plane-dockercfg-gs7dd" Mar 12 00:12:29 crc kubenswrapper[4948]: I0312 00:12:29.795754 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7d9c768c99-hpghc" event={"ID":"ea491ee5-9d95-4a63-896b-4ebd3ceec50a","Type":"ContainerStarted","Data":"42cc9f9fca414ac17ad06628e963a6c0ae151a3454f8ab5491f17784dd6756b6"} Mar 12 00:12:29 crc kubenswrapper[4948]: I0312 00:12:29.795797 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-authentication/oauth-openshift-7d9c768c99-hpghc" event={"ID":"ea491ee5-9d95-4a63-896b-4ebd3ceec50a","Type":"ContainerStarted","Data":"1ab8e07727dac53674e5cc622eb40a80cd30aa6bc7694b8681a384fe773ea6a5"} Mar 12 00:12:29 crc kubenswrapper[4948]: I0312 00:12:29.797259 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-authentication/oauth-openshift-7d9c768c99-hpghc" Mar 12 00:12:29 crc kubenswrapper[4948]: I0312 00:12:29.810864 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-node-metrics-cert" Mar 12 00:12:29 crc kubenswrapper[4948]: I0312 00:12:29.825331 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-authentication/oauth-openshift-7d9c768c99-hpghc" podStartSLOduration=63.82528718 podStartE2EDuration="1m3.82528718s" podCreationTimestamp="2026-03-12 00:11:26 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:12:29.824199725 +0000 UTC m=+349.279803463" watchObservedRunningTime="2026-03-12 00:12:29.82528718 +0000 UTC m=+349.280890958" Mar 12 00:12:29 crc kubenswrapper[4948]: I0312 00:12:29.864925 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-controller-manager"/"config" Mar 12 00:12:29 crc kubenswrapper[4948]: I0312 00:12:29.910475 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-authentication/oauth-openshift-7d9c768c99-hpghc" Mar 12 00:12:29 crc kubenswrapper[4948]: I0312 00:12:29.973343 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver"/"openshift-service-ca.crt" Mar 12 00:12:29 crc kubenswrapper[4948]: I0312 00:12:29.982148 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-scheduler-operator"/"kube-root-ca.crt" Mar 12 00:12:30 crc kubenswrapper[4948]: I0312 00:12:30.161944 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-console"/"default-dockercfg-chnjx" Mar 12 00:12:30 crc kubenswrapper[4948]: I0312 00:12:30.175475 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"openshift-service-ca.crt" Mar 12 00:12:30 crc kubenswrapper[4948]: I0312 00:12:30.230734 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"olm-operator-serving-cert" Mar 12 00:12:30 crc kubenswrapper[4948]: I0312 00:12:30.356637 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-service-ca-bundle" Mar 12 00:12:30 crc kubenswrapper[4948]: I0312 00:12:30.382735 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-script-lib" Mar 12 00:12:30 crc kubenswrapper[4948]: I0312 00:12:30.473386 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"audit-1" Mar 12 00:12:30 crc kubenswrapper[4948]: I0312 00:12:30.504047 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-route-controller-manager"/"openshift-service-ca.crt" Mar 12 00:12:30 crc kubenswrapper[4948]: I0312 00:12:30.509504 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"console-config" Mar 12 00:12:30 crc kubenswrapper[4948]: I0312 00:12:30.543012 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-config-operator"/"machine-config-server-dockercfg-qx5rd" Mar 12 00:12:30 crc kubenswrapper[4948]: I0312 00:12:30.591914 4948 reflector.go:368] Caches populated for *v1.CSIDriver from k8s.io/client-go/informers/factory.go:160 Mar 12 00:12:30 crc kubenswrapper[4948]: I0312 00:12:30.659522 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"openshift-service-ca.crt" Mar 12 00:12:30 crc kubenswrapper[4948]: I0312 00:12:30.716632 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"hostpath-provisioner"/"openshift-service-ca.crt" Mar 12 00:12:30 crc kubenswrapper[4948]: I0312 00:12:30.780723 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-etcd-operator"/"etcd-ca-bundle" Mar 12 00:12:30 crc kubenswrapper[4948]: I0312 00:12:30.875989 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ovn-kubernetes"/"ovnkube-config" Mar 12 00:12:30 crc kubenswrapper[4948]: I0312 00:12:30.953682 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-version"/"openshift-service-ca.crt" Mar 12 00:12:30 crc kubenswrapper[4948]: I0312 00:12:30.999915 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-oauth-apiserver"/"kube-root-ca.crt" Mar 12 00:12:31 crc kubenswrapper[4948]: I0312 00:12:31.010635 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca-operator"/"openshift-service-ca.crt" Mar 12 00:12:31 crc kubenswrapper[4948]: I0312 00:12:31.077682 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"openshift-service-ca.crt" Mar 12 00:12:31 crc kubenswrapper[4948]: I0312 00:12:31.233923 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"metrics-tls" Mar 12 00:12:31 crc kubenswrapper[4948]: I0312 00:12:31.296377 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver"/"etcd-client" Mar 12 00:12:31 crc kubenswrapper[4948]: I0312 00:12:31.338239 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-network-node-identity"/"kube-root-ca.crt" Mar 12 00:12:31 crc kubenswrapper[4948]: I0312 00:12:31.412455 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-service-ca"/"openshift-service-ca.crt" Mar 12 00:12:31 crc kubenswrapper[4948]: I0312 00:12:31.421250 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-kube-storage-version-migrator-operator"/"openshift-service-ca.crt" Mar 12 00:12:31 crc kubenswrapper[4948]: I0312 00:12:31.479369 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-canary"/"canary-serving-cert" Mar 12 00:12:31 crc kubenswrapper[4948]: I0312 00:12:31.510533 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress"/"service-ca-bundle" Mar 12 00:12:31 crc kubenswrapper[4948]: I0312 00:12:31.549076 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"kube-root-ca.crt" Mar 12 00:12:31 crc kubenswrapper[4948]: I0312 00:12:31.553630 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-apiserver-operator"/"openshift-apiserver-operator-dockercfg-xtcjv" Mar 12 00:12:31 crc kubenswrapper[4948]: I0312 00:12:31.568319 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-image-registry"/"cluster-image-registry-operator-dockercfg-m4qtx" Mar 12 00:12:31 crc kubenswrapper[4948]: I0312 00:12:31.769790 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-ingress-canary"/"kube-root-ca.crt" Mar 12 00:12:32 crc kubenswrapper[4948]: I0312 00:12:32.194449 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-machine-config-operator"/"openshift-service-ca.crt" Mar 12 00:12:32 crc kubenswrapper[4948]: I0312 00:12:32.196802 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"kube-root-ca.crt" Mar 12 00:12:32 crc kubenswrapper[4948]: I0312 00:12:32.259540 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"trusted-ca-bundle" Mar 12 00:12:32 crc kubenswrapper[4948]: I0312 00:12:32.485016 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-multus"/"kube-root-ca.crt" Mar 12 00:12:32 crc kubenswrapper[4948]: I0312 00:12:32.538945 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress-operator"/"ingress-operator-dockercfg-7lnqk" Mar 12 00:12:32 crc kubenswrapper[4948]: I0312 00:12:32.680468 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-authentication-operator"/"openshift-service-ca.crt" Mar 12 00:12:32 crc kubenswrapper[4948]: I0312 00:12:32.715823 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"marketplace-operator-dockercfg-5nsgg" Mar 12 00:12:32 crc kubenswrapper[4948]: I0312 00:12:32.899204 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-service-ca-operator"/"service-ca-operator-dockercfg-rg9jl" Mar 12 00:12:32 crc kubenswrapper[4948]: I0312 00:12:32.915271 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-marketplace"/"openshift-service-ca.crt" Mar 12 00:12:33 crc kubenswrapper[4948]: I0312 00:12:33.066658 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-apiserver-operator"/"kube-root-ca.crt" Mar 12 00:12:33 crc kubenswrapper[4948]: I0312 00:12:33.143761 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ovn-kubernetes"/"ovn-kubernetes-node-dockercfg-pwtwl" Mar 12 00:12:33 crc kubenswrapper[4948]: I0312 00:12:33.145273 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-cluster-machine-approver"/"openshift-service-ca.crt" Mar 12 00:12:33 crc kubenswrapper[4948]: I0312 00:12:33.375285 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 12 00:12:33 crc kubenswrapper[4948]: I0312 00:12:33.658840 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-machine-api"/"machine-api-operator-dockercfg-mfbb7" Mar 12 00:12:33 crc kubenswrapper[4948]: I0312 00:12:33.920683 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-kube-storage-version-migrator"/"kube-storage-version-migrator-sa-dockercfg-5xfcg" Mar 12 00:12:33 crc kubenswrapper[4948]: I0312 00:12:33.926131 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-ingress"/"router-certs-default" Mar 12 00:12:34 crc kubenswrapper[4948]: I0312 00:12:34.260524 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"catalog-operator-serving-cert" Mar 12 00:12:34 crc kubenswrapper[4948]: I0312 00:12:34.612989 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-console"/"openshift-service-ca.crt" Mar 12 00:12:34 crc kubenswrapper[4948]: I0312 00:12:34.707544 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Mar 12 00:12:34 crc kubenswrapper[4948]: I0312 00:12:34.707632 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 12 00:12:34 crc kubenswrapper[4948]: I0312 00:12:34.759029 4948 reflector.go:368] Caches populated for *v1.RuntimeClass from k8s.io/client-go/informers/factory.go:160 Mar 12 00:12:34 crc kubenswrapper[4948]: I0312 00:12:34.784037 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 12 00:12:34 crc kubenswrapper[4948]: I0312 00:12:34.784127 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 12 00:12:34 crc kubenswrapper[4948]: I0312 00:12:34.784173 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 12 00:12:34 crc kubenswrapper[4948]: I0312 00:12:34.784249 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 12 00:12:34 crc kubenswrapper[4948]: I0312 00:12:34.784311 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") pod \"f85e55b1a89d02b0cb034b1ea31ed45a\" (UID: \"f85e55b1a89d02b0cb034b1ea31ed45a\") " Mar 12 00:12:34 crc kubenswrapper[4948]: I0312 00:12:34.784345 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log" (OuterVolumeSpecName: "var-log") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 00:12:34 crc kubenswrapper[4948]: I0312 00:12:34.784344 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir" (OuterVolumeSpecName: "resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 00:12:34 crc kubenswrapper[4948]: I0312 00:12:34.784400 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock" (OuterVolumeSpecName: "var-lock") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "var-lock". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 00:12:34 crc kubenswrapper[4948]: I0312 00:12:34.784555 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests" (OuterVolumeSpecName: "manifests") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "manifests". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 00:12:34 crc kubenswrapper[4948]: I0312 00:12:34.784729 4948 reconciler_common.go:293] "Volume detached for volume \"resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 12 00:12:34 crc kubenswrapper[4948]: I0312 00:12:34.784751 4948 reconciler_common.go:293] "Volume detached for volume \"var-log\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-log\") on node \"crc\" DevicePath \"\"" Mar 12 00:12:34 crc kubenswrapper[4948]: I0312 00:12:34.784767 4948 reconciler_common.go:293] "Volume detached for volume \"var-lock\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-var-lock\") on node \"crc\" DevicePath \"\"" Mar 12 00:12:34 crc kubenswrapper[4948]: I0312 00:12:34.784784 4948 reconciler_common.go:293] "Volume detached for volume \"manifests\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-manifests\") on node \"crc\" DevicePath \"\"" Mar 12 00:12:34 crc kubenswrapper[4948]: I0312 00:12:34.801666 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir" (OuterVolumeSpecName: "pod-resource-dir") pod "f85e55b1a89d02b0cb034b1ea31ed45a" (UID: "f85e55b1a89d02b0cb034b1ea31ed45a"). InnerVolumeSpecName "pod-resource-dir". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 00:12:34 crc kubenswrapper[4948]: I0312 00:12:34.829230 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-apiserver_kube-apiserver-startup-monitor-crc_f85e55b1a89d02b0cb034b1ea31ed45a/startup-monitor/0.log" Mar 12 00:12:34 crc kubenswrapper[4948]: I0312 00:12:34.829289 4948 generic.go:334] "Generic (PLEG): container finished" podID="f85e55b1a89d02b0cb034b1ea31ed45a" containerID="b8f42af8d5706c672759fe6b3ea65eaccb27bc3e80ed254e63660d83b69bfcde" exitCode=137 Mar 12 00:12:34 crc kubenswrapper[4948]: I0312 00:12:34.829376 4948 scope.go:117] "RemoveContainer" containerID="b8f42af8d5706c672759fe6b3ea65eaccb27bc3e80ed254e63660d83b69bfcde" Mar 12 00:12:34 crc kubenswrapper[4948]: I0312 00:12:34.829453 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-kube-apiserver/kube-apiserver-startup-monitor-crc" Mar 12 00:12:34 crc kubenswrapper[4948]: I0312 00:12:34.859235 4948 scope.go:117] "RemoveContainer" containerID="b8f42af8d5706c672759fe6b3ea65eaccb27bc3e80ed254e63660d83b69bfcde" Mar 12 00:12:34 crc kubenswrapper[4948]: E0312 00:12:34.859881 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b8f42af8d5706c672759fe6b3ea65eaccb27bc3e80ed254e63660d83b69bfcde\": container with ID starting with b8f42af8d5706c672759fe6b3ea65eaccb27bc3e80ed254e63660d83b69bfcde not found: ID does not exist" containerID="b8f42af8d5706c672759fe6b3ea65eaccb27bc3e80ed254e63660d83b69bfcde" Mar 12 00:12:34 crc kubenswrapper[4948]: I0312 00:12:34.859945 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b8f42af8d5706c672759fe6b3ea65eaccb27bc3e80ed254e63660d83b69bfcde"} err="failed to get container status \"b8f42af8d5706c672759fe6b3ea65eaccb27bc3e80ed254e63660d83b69bfcde\": rpc error: code = NotFound desc = could not find container \"b8f42af8d5706c672759fe6b3ea65eaccb27bc3e80ed254e63660d83b69bfcde\": container with ID starting with b8f42af8d5706c672759fe6b3ea65eaccb27bc3e80ed254e63660d83b69bfcde not found: ID does not exist" Mar 12 00:12:34 crc kubenswrapper[4948]: I0312 00:12:34.886601 4948 reconciler_common.go:293] "Volume detached for volume \"pod-resource-dir\" (UniqueName: \"kubernetes.io/host-path/f85e55b1a89d02b0cb034b1ea31ed45a-pod-resource-dir\") on node \"crc\" DevicePath \"\"" Mar 12 00:12:35 crc kubenswrapper[4948]: I0312 00:12:35.322706 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" path="/var/lib/kubelet/pods/f85e55b1a89d02b0cb034b1ea31ed45a/volumes" Mar 12 00:12:53 crc kubenswrapper[4948]: I0312 00:12:53.962977 4948 generic.go:334] "Generic (PLEG): container finished" podID="db40ba09-a72f-43d5-8037-e71c96e2bf5b" containerID="77256205eaf26927b82537a261f9e9f9e55832d90d862711c36f6862edd98955" exitCode=0 Mar 12 00:12:53 crc kubenswrapper[4948]: I0312 00:12:53.963599 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-zkbhv" event={"ID":"db40ba09-a72f-43d5-8037-e71c96e2bf5b","Type":"ContainerDied","Data":"77256205eaf26927b82537a261f9e9f9e55832d90d862711c36f6862edd98955"} Mar 12 00:12:53 crc kubenswrapper[4948]: I0312 00:12:53.964451 4948 scope.go:117] "RemoveContainer" containerID="77256205eaf26927b82537a261f9e9f9e55832d90d862711c36f6862edd98955" Mar 12 00:12:53 crc kubenswrapper[4948]: I0312 00:12:53.970975 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Mar 12 00:12:53 crc kubenswrapper[4948]: I0312 00:12:53.974902 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 12 00:12:53 crc kubenswrapper[4948]: I0312 00:12:53.975680 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/0.log" Mar 12 00:12:53 crc kubenswrapper[4948]: I0312 00:12:53.975727 4948 generic.go:334] "Generic (PLEG): container finished" podID="f614b9022728cf315e60c057852e563e" containerID="6d4ce0622c9e3a674aaaf793d0d56390132a140e51a9c7f4d37a94db9dce816f" exitCode=137 Mar 12 00:12:53 crc kubenswrapper[4948]: I0312 00:12:53.975761 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerDied","Data":"6d4ce0622c9e3a674aaaf793d0d56390132a140e51a9c7f4d37a94db9dce816f"} Mar 12 00:12:53 crc kubenswrapper[4948]: I0312 00:12:53.975803 4948 scope.go:117] "RemoveContainer" containerID="fb45482a404d6696cfbc3919997e53146b7b9dd1536056a55374574ae01ff193" Mar 12 00:12:54 crc kubenswrapper[4948]: I0312 00:12:54.986583 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-zkbhv" event={"ID":"db40ba09-a72f-43d5-8037-e71c96e2bf5b","Type":"ContainerStarted","Data":"6c01edba31d8a4b8c56434839bfda737c5034ad32ad39b244d484ce89accf516"} Mar 12 00:12:54 crc kubenswrapper[4948]: I0312 00:12:54.987363 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-zkbhv" Mar 12 00:12:54 crc kubenswrapper[4948]: I0312 00:12:54.992038 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-zkbhv" Mar 12 00:12:54 crc kubenswrapper[4948]: I0312 00:12:54.996099 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/kube-controller-manager/1.log" Mar 12 00:12:54 crc kubenswrapper[4948]: I0312 00:12:54.999452 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-kube-controller-manager_kube-controller-manager-crc_f614b9022728cf315e60c057852e563e/cluster-policy-controller/0.log" Mar 12 00:12:54 crc kubenswrapper[4948]: I0312 00:12:54.999567 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-kube-controller-manager/kube-controller-manager-crc" event={"ID":"f614b9022728cf315e60c057852e563e","Type":"ContainerStarted","Data":"0bba018852ee48c3ed1fbec6e49e833accb909a9366e22409ae7bd00f2f6e584"} Mar 12 00:13:01 crc kubenswrapper[4948]: I0312 00:13:01.878478 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 12 00:13:03 crc kubenswrapper[4948]: I0312 00:13:03.220840 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 12 00:13:03 crc kubenswrapper[4948]: I0312 00:13:03.229642 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 12 00:13:04 crc kubenswrapper[4948]: I0312 00:13:04.069161 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-kube-controller-manager/kube-controller-manager-crc" Mar 12 00:13:09 crc kubenswrapper[4948]: I0312 00:13:09.952812 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29554572-szktg"] Mar 12 00:13:09 crc kubenswrapper[4948]: E0312 00:13:09.953519 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 12 00:13:09 crc kubenswrapper[4948]: I0312 00:13:09.953531 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 12 00:13:09 crc kubenswrapper[4948]: I0312 00:13:09.953629 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="f85e55b1a89d02b0cb034b1ea31ed45a" containerName="startup-monitor" Mar 12 00:13:09 crc kubenswrapper[4948]: I0312 00:13:09.954027 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29554572-szktg" Mar 12 00:13:09 crc kubenswrapper[4948]: I0312 00:13:09.956635 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-44g5m" Mar 12 00:13:09 crc kubenswrapper[4948]: I0312 00:13:09.956911 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 00:13:09 crc kubenswrapper[4948]: I0312 00:13:09.957578 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 00:13:09 crc kubenswrapper[4948]: I0312 00:13:09.967570 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29554572-szktg"] Mar 12 00:13:10 crc kubenswrapper[4948]: I0312 00:13:10.145522 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4ffqs\" (UniqueName: \"kubernetes.io/projected/7401d71e-946e-4b13-b187-13db4b1be35b-kube-api-access-4ffqs\") pod \"auto-csr-approver-29554572-szktg\" (UID: \"7401d71e-946e-4b13-b187-13db4b1be35b\") " pod="openshift-infra/auto-csr-approver-29554572-szktg" Mar 12 00:13:10 crc kubenswrapper[4948]: I0312 00:13:10.246684 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4ffqs\" (UniqueName: \"kubernetes.io/projected/7401d71e-946e-4b13-b187-13db4b1be35b-kube-api-access-4ffqs\") pod \"auto-csr-approver-29554572-szktg\" (UID: \"7401d71e-946e-4b13-b187-13db4b1be35b\") " pod="openshift-infra/auto-csr-approver-29554572-szktg" Mar 12 00:13:10 crc kubenswrapper[4948]: I0312 00:13:10.277367 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4ffqs\" (UniqueName: \"kubernetes.io/projected/7401d71e-946e-4b13-b187-13db4b1be35b-kube-api-access-4ffqs\") pod \"auto-csr-approver-29554572-szktg\" (UID: \"7401d71e-946e-4b13-b187-13db4b1be35b\") " pod="openshift-infra/auto-csr-approver-29554572-szktg" Mar 12 00:13:10 crc kubenswrapper[4948]: I0312 00:13:10.571983 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29554572-szktg" Mar 12 00:13:10 crc kubenswrapper[4948]: I0312 00:13:10.974053 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29554572-szktg"] Mar 12 00:13:11 crc kubenswrapper[4948]: I0312 00:13:11.108083 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29554572-szktg" event={"ID":"7401d71e-946e-4b13-b187-13db4b1be35b","Type":"ContainerStarted","Data":"be3fa3b545746b7fa15c6b24d2ac86b73a2eed484442f76cb0122e74ecf2fc70"} Mar 12 00:13:13 crc kubenswrapper[4948]: I0312 00:13:13.121736 4948 generic.go:334] "Generic (PLEG): container finished" podID="7401d71e-946e-4b13-b187-13db4b1be35b" containerID="24742ed07ef4b5fa12b73123be5bed0fe00189424e312e3a45057377d14859df" exitCode=0 Mar 12 00:13:13 crc kubenswrapper[4948]: I0312 00:13:13.121811 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29554572-szktg" event={"ID":"7401d71e-946e-4b13-b187-13db4b1be35b","Type":"ContainerDied","Data":"24742ed07ef4b5fa12b73123be5bed0fe00189424e312e3a45057377d14859df"} Mar 12 00:13:14 crc kubenswrapper[4948]: I0312 00:13:14.402774 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29554572-szktg" Mar 12 00:13:14 crc kubenswrapper[4948]: I0312 00:13:14.597050 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4ffqs\" (UniqueName: \"kubernetes.io/projected/7401d71e-946e-4b13-b187-13db4b1be35b-kube-api-access-4ffqs\") pod \"7401d71e-946e-4b13-b187-13db4b1be35b\" (UID: \"7401d71e-946e-4b13-b187-13db4b1be35b\") " Mar 12 00:13:14 crc kubenswrapper[4948]: I0312 00:13:14.604258 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/7401d71e-946e-4b13-b187-13db4b1be35b-kube-api-access-4ffqs" (OuterVolumeSpecName: "kube-api-access-4ffqs") pod "7401d71e-946e-4b13-b187-13db4b1be35b" (UID: "7401d71e-946e-4b13-b187-13db4b1be35b"). InnerVolumeSpecName "kube-api-access-4ffqs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:13:14 crc kubenswrapper[4948]: I0312 00:13:14.698556 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4ffqs\" (UniqueName: \"kubernetes.io/projected/7401d71e-946e-4b13-b187-13db4b1be35b-kube-api-access-4ffqs\") on node \"crc\" DevicePath \"\"" Mar 12 00:13:15 crc kubenswrapper[4948]: I0312 00:13:15.140435 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29554572-szktg" Mar 12 00:13:15 crc kubenswrapper[4948]: I0312 00:13:15.140823 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29554572-szktg" event={"ID":"7401d71e-946e-4b13-b187-13db4b1be35b","Type":"ContainerDied","Data":"be3fa3b545746b7fa15c6b24d2ac86b73a2eed484442f76cb0122e74ecf2fc70"} Mar 12 00:13:15 crc kubenswrapper[4948]: I0312 00:13:15.140872 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="be3fa3b545746b7fa15c6b24d2ac86b73a2eed484442f76cb0122e74ecf2fc70" Mar 12 00:13:44 crc kubenswrapper[4948]: I0312 00:13:44.784966 4948 patch_prober.go:28] interesting pod/machine-config-daemon-m4xwc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 00:13:44 crc kubenswrapper[4948]: I0312 00:13:44.787358 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" podUID="7e483c2b-08f0-4e92-8e4a-b7281f30af3e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.023908 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rflhd"] Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.025163 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-rflhd" podUID="5184125a-dde1-4db4-9d6b-5b3c45d8d82a" containerName="registry-server" containerID="cri-o://bf689a8e9a12667b22bba99be51be88469d4e2cd090bfa6efe2c69a940fe8322" gracePeriod=30 Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.039451 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-v8h88"] Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.040380 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-v8h88" podUID="5d83d568-3968-4ea6-b548-511c951defa5" containerName="registry-server" containerID="cri-o://05fd166aeeb70277dab1df229c8cbde7b85139a276d0732903b8b28713344011" gracePeriod=30 Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.049047 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-zkbhv"] Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.049384 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/marketplace-operator-79b997595-zkbhv" podUID="db40ba09-a72f-43d5-8037-e71c96e2bf5b" containerName="marketplace-operator" containerID="cri-o://6c01edba31d8a4b8c56434839bfda737c5034ad32ad39b244d484ce89accf516" gracePeriod=30 Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.060468 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zc9zh"] Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.060922 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-zc9zh" podUID="f8cf7ee1-e662-4afd-9e08-112171677c22" containerName="registry-server" containerID="cri-o://326ebf4817bc847d36148437a6dcda1508f716edd2d55a3be608cafdfbfb5f7e" gracePeriod=30 Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.069366 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-btd4j"] Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.069666 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-btd4j" podUID="8a6ce79a-afdb-4e21-9e24-8de6924279bf" containerName="registry-server" containerID="cri-o://b388539342b4cfeef5eebb31026e89abad176dae01f1b209244ea76136260c19" gracePeriod=30 Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.072977 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5k5ph"] Mar 12 00:13:51 crc kubenswrapper[4948]: E0312 00:13:51.073227 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="7401d71e-946e-4b13-b187-13db4b1be35b" containerName="oc" Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.073248 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="7401d71e-946e-4b13-b187-13db4b1be35b" containerName="oc" Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.073403 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="7401d71e-946e-4b13-b187-13db4b1be35b" containerName="oc" Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.073857 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-5k5ph" Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.089638 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5k5ph"] Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.186343 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d39a6a83-8fa5-47e4-8047-4b4e713e1a64-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-5k5ph\" (UID: \"d39a6a83-8fa5-47e4-8047-4b4e713e1a64\") " pod="openshift-marketplace/marketplace-operator-79b997595-5k5ph" Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.186410 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-tbrx2\" (UniqueName: \"kubernetes.io/projected/d39a6a83-8fa5-47e4-8047-4b4e713e1a64-kube-api-access-tbrx2\") pod \"marketplace-operator-79b997595-5k5ph\" (UID: \"d39a6a83-8fa5-47e4-8047-4b4e713e1a64\") " pod="openshift-marketplace/marketplace-operator-79b997595-5k5ph" Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.186449 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d39a6a83-8fa5-47e4-8047-4b4e713e1a64-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-5k5ph\" (UID: \"d39a6a83-8fa5-47e4-8047-4b4e713e1a64\") " pod="openshift-marketplace/marketplace-operator-79b997595-5k5ph" Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.287970 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d39a6a83-8fa5-47e4-8047-4b4e713e1a64-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-5k5ph\" (UID: \"d39a6a83-8fa5-47e4-8047-4b4e713e1a64\") " pod="openshift-marketplace/marketplace-operator-79b997595-5k5ph" Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.288033 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d39a6a83-8fa5-47e4-8047-4b4e713e1a64-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-5k5ph\" (UID: \"d39a6a83-8fa5-47e4-8047-4b4e713e1a64\") " pod="openshift-marketplace/marketplace-operator-79b997595-5k5ph" Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.288073 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-tbrx2\" (UniqueName: \"kubernetes.io/projected/d39a6a83-8fa5-47e4-8047-4b4e713e1a64-kube-api-access-tbrx2\") pod \"marketplace-operator-79b997595-5k5ph\" (UID: \"d39a6a83-8fa5-47e4-8047-4b4e713e1a64\") " pod="openshift-marketplace/marketplace-operator-79b997595-5k5ph" Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.289482 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/d39a6a83-8fa5-47e4-8047-4b4e713e1a64-marketplace-trusted-ca\") pod \"marketplace-operator-79b997595-5k5ph\" (UID: \"d39a6a83-8fa5-47e4-8047-4b4e713e1a64\") " pod="openshift-marketplace/marketplace-operator-79b997595-5k5ph" Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.294642 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/d39a6a83-8fa5-47e4-8047-4b4e713e1a64-marketplace-operator-metrics\") pod \"marketplace-operator-79b997595-5k5ph\" (UID: \"d39a6a83-8fa5-47e4-8047-4b4e713e1a64\") " pod="openshift-marketplace/marketplace-operator-79b997595-5k5ph" Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.313981 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-tbrx2\" (UniqueName: \"kubernetes.io/projected/d39a6a83-8fa5-47e4-8047-4b4e713e1a64-kube-api-access-tbrx2\") pod \"marketplace-operator-79b997595-5k5ph\" (UID: \"d39a6a83-8fa5-47e4-8047-4b4e713e1a64\") " pod="openshift-marketplace/marketplace-operator-79b997595-5k5ph" Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.392950 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-5k5ph" Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.431627 4948 generic.go:334] "Generic (PLEG): container finished" podID="5184125a-dde1-4db4-9d6b-5b3c45d8d82a" containerID="bf689a8e9a12667b22bba99be51be88469d4e2cd090bfa6efe2c69a940fe8322" exitCode=0 Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.431679 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rflhd" event={"ID":"5184125a-dde1-4db4-9d6b-5b3c45d8d82a","Type":"ContainerDied","Data":"bf689a8e9a12667b22bba99be51be88469d4e2cd090bfa6efe2c69a940fe8322"} Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.433534 4948 generic.go:334] "Generic (PLEG): container finished" podID="f8cf7ee1-e662-4afd-9e08-112171677c22" containerID="326ebf4817bc847d36148437a6dcda1508f716edd2d55a3be608cafdfbfb5f7e" exitCode=0 Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.433572 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zc9zh" event={"ID":"f8cf7ee1-e662-4afd-9e08-112171677c22","Type":"ContainerDied","Data":"326ebf4817bc847d36148437a6dcda1508f716edd2d55a3be608cafdfbfb5f7e"} Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.435217 4948 generic.go:334] "Generic (PLEG): container finished" podID="8a6ce79a-afdb-4e21-9e24-8de6924279bf" containerID="b388539342b4cfeef5eebb31026e89abad176dae01f1b209244ea76136260c19" exitCode=0 Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.435253 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-btd4j" event={"ID":"8a6ce79a-afdb-4e21-9e24-8de6924279bf","Type":"ContainerDied","Data":"b388539342b4cfeef5eebb31026e89abad176dae01f1b209244ea76136260c19"} Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.442554 4948 generic.go:334] "Generic (PLEG): container finished" podID="5d83d568-3968-4ea6-b548-511c951defa5" containerID="05fd166aeeb70277dab1df229c8cbde7b85139a276d0732903b8b28713344011" exitCode=0 Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.442623 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v8h88" event={"ID":"5d83d568-3968-4ea6-b548-511c951defa5","Type":"ContainerDied","Data":"05fd166aeeb70277dab1df229c8cbde7b85139a276d0732903b8b28713344011"} Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.442654 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-v8h88" event={"ID":"5d83d568-3968-4ea6-b548-511c951defa5","Type":"ContainerDied","Data":"a6dc12e7d1d52b0588ce8348ac2a6750d029729d64b7ddce4efc6ab839856f2e"} Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.442664 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a6dc12e7d1d52b0588ce8348ac2a6750d029729d64b7ddce4efc6ab839856f2e" Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.445101 4948 generic.go:334] "Generic (PLEG): container finished" podID="db40ba09-a72f-43d5-8037-e71c96e2bf5b" containerID="6c01edba31d8a4b8c56434839bfda737c5034ad32ad39b244d484ce89accf516" exitCode=0 Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.445125 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-zkbhv" event={"ID":"db40ba09-a72f-43d5-8037-e71c96e2bf5b","Type":"ContainerDied","Data":"6c01edba31d8a4b8c56434839bfda737c5034ad32ad39b244d484ce89accf516"} Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.445147 4948 scope.go:117] "RemoveContainer" containerID="77256205eaf26927b82537a261f9e9f9e55832d90d862711c36f6862edd98955" Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.449743 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v8h88" Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.451287 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rflhd" Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.456115 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zc9zh" Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.469803 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-zkbhv" Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.520115 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-btd4j" Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.590480 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5d83d568-3968-4ea6-b548-511c951defa5-utilities\") pod \"5d83d568-3968-4ea6-b548-511c951defa5\" (UID: \"5d83d568-3968-4ea6-b548-511c951defa5\") " Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.590523 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/db40ba09-a72f-43d5-8037-e71c96e2bf5b-marketplace-trusted-ca\") pod \"db40ba09-a72f-43d5-8037-e71c96e2bf5b\" (UID: \"db40ba09-a72f-43d5-8037-e71c96e2bf5b\") " Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.590569 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/db40ba09-a72f-43d5-8037-e71c96e2bf5b-marketplace-operator-metrics\") pod \"db40ba09-a72f-43d5-8037-e71c96e2bf5b\" (UID: \"db40ba09-a72f-43d5-8037-e71c96e2bf5b\") " Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.590599 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-tck76\" (UniqueName: \"kubernetes.io/projected/f8cf7ee1-e662-4afd-9e08-112171677c22-kube-api-access-tck76\") pod \"f8cf7ee1-e662-4afd-9e08-112171677c22\" (UID: \"f8cf7ee1-e662-4afd-9e08-112171677c22\") " Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.590629 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-lbkcx\" (UniqueName: \"kubernetes.io/projected/5d83d568-3968-4ea6-b548-511c951defa5-kube-api-access-lbkcx\") pod \"5d83d568-3968-4ea6-b548-511c951defa5\" (UID: \"5d83d568-3968-4ea6-b548-511c951defa5\") " Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.590649 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5184125a-dde1-4db4-9d6b-5b3c45d8d82a-catalog-content\") pod \"5184125a-dde1-4db4-9d6b-5b3c45d8d82a\" (UID: \"5184125a-dde1-4db4-9d6b-5b3c45d8d82a\") " Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.590667 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rljd9\" (UniqueName: \"kubernetes.io/projected/5184125a-dde1-4db4-9d6b-5b3c45d8d82a-kube-api-access-rljd9\") pod \"5184125a-dde1-4db4-9d6b-5b3c45d8d82a\" (UID: \"5184125a-dde1-4db4-9d6b-5b3c45d8d82a\") " Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.590688 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5d83d568-3968-4ea6-b548-511c951defa5-catalog-content\") pod \"5d83d568-3968-4ea6-b548-511c951defa5\" (UID: \"5d83d568-3968-4ea6-b548-511c951defa5\") " Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.590703 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8cf7ee1-e662-4afd-9e08-112171677c22-utilities\") pod \"f8cf7ee1-e662-4afd-9e08-112171677c22\" (UID: \"f8cf7ee1-e662-4afd-9e08-112171677c22\") " Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.590723 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-g7gzh\" (UniqueName: \"kubernetes.io/projected/db40ba09-a72f-43d5-8037-e71c96e2bf5b-kube-api-access-g7gzh\") pod \"db40ba09-a72f-43d5-8037-e71c96e2bf5b\" (UID: \"db40ba09-a72f-43d5-8037-e71c96e2bf5b\") " Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.590758 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8cf7ee1-e662-4afd-9e08-112171677c22-catalog-content\") pod \"f8cf7ee1-e662-4afd-9e08-112171677c22\" (UID: \"f8cf7ee1-e662-4afd-9e08-112171677c22\") " Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.590774 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5184125a-dde1-4db4-9d6b-5b3c45d8d82a-utilities\") pod \"5184125a-dde1-4db4-9d6b-5b3c45d8d82a\" (UID: \"5184125a-dde1-4db4-9d6b-5b3c45d8d82a\") " Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.591452 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5d83d568-3968-4ea6-b548-511c951defa5-utilities" (OuterVolumeSpecName: "utilities") pod "5d83d568-3968-4ea6-b548-511c951defa5" (UID: "5d83d568-3968-4ea6-b548-511c951defa5"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.591536 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5184125a-dde1-4db4-9d6b-5b3c45d8d82a-utilities" (OuterVolumeSpecName: "utilities") pod "5184125a-dde1-4db4-9d6b-5b3c45d8d82a" (UID: "5184125a-dde1-4db4-9d6b-5b3c45d8d82a"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.591841 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8cf7ee1-e662-4afd-9e08-112171677c22-utilities" (OuterVolumeSpecName: "utilities") pod "f8cf7ee1-e662-4afd-9e08-112171677c22" (UID: "f8cf7ee1-e662-4afd-9e08-112171677c22"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.595135 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5d83d568-3968-4ea6-b548-511c951defa5-kube-api-access-lbkcx" (OuterVolumeSpecName: "kube-api-access-lbkcx") pod "5d83d568-3968-4ea6-b548-511c951defa5" (UID: "5d83d568-3968-4ea6-b548-511c951defa5"). InnerVolumeSpecName "kube-api-access-lbkcx". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.595185 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5184125a-dde1-4db4-9d6b-5b3c45d8d82a-kube-api-access-rljd9" (OuterVolumeSpecName: "kube-api-access-rljd9") pod "5184125a-dde1-4db4-9d6b-5b3c45d8d82a" (UID: "5184125a-dde1-4db4-9d6b-5b3c45d8d82a"). InnerVolumeSpecName "kube-api-access-rljd9". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.595262 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f8cf7ee1-e662-4afd-9e08-112171677c22-kube-api-access-tck76" (OuterVolumeSpecName: "kube-api-access-tck76") pod "f8cf7ee1-e662-4afd-9e08-112171677c22" (UID: "f8cf7ee1-e662-4afd-9e08-112171677c22"). InnerVolumeSpecName "kube-api-access-tck76". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.595723 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/db40ba09-a72f-43d5-8037-e71c96e2bf5b-marketplace-operator-metrics" (OuterVolumeSpecName: "marketplace-operator-metrics") pod "db40ba09-a72f-43d5-8037-e71c96e2bf5b" (UID: "db40ba09-a72f-43d5-8037-e71c96e2bf5b"). InnerVolumeSpecName "marketplace-operator-metrics". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.607528 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/db40ba09-a72f-43d5-8037-e71c96e2bf5b-marketplace-trusted-ca" (OuterVolumeSpecName: "marketplace-trusted-ca") pod "db40ba09-a72f-43d5-8037-e71c96e2bf5b" (UID: "db40ba09-a72f-43d5-8037-e71c96e2bf5b"). InnerVolumeSpecName "marketplace-trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.615074 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/db40ba09-a72f-43d5-8037-e71c96e2bf5b-kube-api-access-g7gzh" (OuterVolumeSpecName: "kube-api-access-g7gzh") pod "db40ba09-a72f-43d5-8037-e71c96e2bf5b" (UID: "db40ba09-a72f-43d5-8037-e71c96e2bf5b"). InnerVolumeSpecName "kube-api-access-g7gzh". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.616846 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/f8cf7ee1-e662-4afd-9e08-112171677c22-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "f8cf7ee1-e662-4afd-9e08-112171677c22" (UID: "f8cf7ee1-e662-4afd-9e08-112171677c22"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.652384 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5184125a-dde1-4db4-9d6b-5b3c45d8d82a-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5184125a-dde1-4db4-9d6b-5b3c45d8d82a" (UID: "5184125a-dde1-4db4-9d6b-5b3c45d8d82a"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.659140 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/5d83d568-3968-4ea6-b548-511c951defa5-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "5d83d568-3968-4ea6-b548-511c951defa5" (UID: "5d83d568-3968-4ea6-b548-511c951defa5"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.691839 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a6ce79a-afdb-4e21-9e24-8de6924279bf-catalog-content\") pod \"8a6ce79a-afdb-4e21-9e24-8de6924279bf\" (UID: \"8a6ce79a-afdb-4e21-9e24-8de6924279bf\") " Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.691962 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a6ce79a-afdb-4e21-9e24-8de6924279bf-utilities\") pod \"8a6ce79a-afdb-4e21-9e24-8de6924279bf\" (UID: \"8a6ce79a-afdb-4e21-9e24-8de6924279bf\") " Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.691984 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-l7p47\" (UniqueName: \"kubernetes.io/projected/8a6ce79a-afdb-4e21-9e24-8de6924279bf-kube-api-access-l7p47\") pod \"8a6ce79a-afdb-4e21-9e24-8de6924279bf\" (UID: \"8a6ce79a-afdb-4e21-9e24-8de6924279bf\") " Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.692171 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-tck76\" (UniqueName: \"kubernetes.io/projected/f8cf7ee1-e662-4afd-9e08-112171677c22-kube-api-access-tck76\") on node \"crc\" DevicePath \"\"" Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.692183 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-lbkcx\" (UniqueName: \"kubernetes.io/projected/5d83d568-3968-4ea6-b548-511c951defa5-kube-api-access-lbkcx\") on node \"crc\" DevicePath \"\"" Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.692191 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5184125a-dde1-4db4-9d6b-5b3c45d8d82a-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.692199 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rljd9\" (UniqueName: \"kubernetes.io/projected/5184125a-dde1-4db4-9d6b-5b3c45d8d82a-kube-api-access-rljd9\") on node \"crc\" DevicePath \"\"" Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.692207 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/5d83d568-3968-4ea6-b548-511c951defa5-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.692216 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/f8cf7ee1-e662-4afd-9e08-112171677c22-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.692224 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-g7gzh\" (UniqueName: \"kubernetes.io/projected/db40ba09-a72f-43d5-8037-e71c96e2bf5b-kube-api-access-g7gzh\") on node \"crc\" DevicePath \"\"" Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.692231 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/f8cf7ee1-e662-4afd-9e08-112171677c22-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.692239 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5184125a-dde1-4db4-9d6b-5b3c45d8d82a-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.692246 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/5d83d568-3968-4ea6-b548-511c951defa5-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.692254 4948 reconciler_common.go:293] "Volume detached for volume \"marketplace-trusted-ca\" (UniqueName: \"kubernetes.io/configmap/db40ba09-a72f-43d5-8037-e71c96e2bf5b-marketplace-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.692264 4948 reconciler_common.go:293] "Volume detached for volume \"marketplace-operator-metrics\" (UniqueName: \"kubernetes.io/secret/db40ba09-a72f-43d5-8037-e71c96e2bf5b-marketplace-operator-metrics\") on node \"crc\" DevicePath \"\"" Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.693255 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a6ce79a-afdb-4e21-9e24-8de6924279bf-utilities" (OuterVolumeSpecName: "utilities") pod "8a6ce79a-afdb-4e21-9e24-8de6924279bf" (UID: "8a6ce79a-afdb-4e21-9e24-8de6924279bf"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.696057 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8a6ce79a-afdb-4e21-9e24-8de6924279bf-kube-api-access-l7p47" (OuterVolumeSpecName: "kube-api-access-l7p47") pod "8a6ce79a-afdb-4e21-9e24-8de6924279bf" (UID: "8a6ce79a-afdb-4e21-9e24-8de6924279bf"). InnerVolumeSpecName "kube-api-access-l7p47". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.793198 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/8a6ce79a-afdb-4e21-9e24-8de6924279bf-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.793231 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-l7p47\" (UniqueName: \"kubernetes.io/projected/8a6ce79a-afdb-4e21-9e24-8de6924279bf-kube-api-access-l7p47\") on node \"crc\" DevicePath \"\"" Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.814351 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/8a6ce79a-afdb-4e21-9e24-8de6924279bf-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "8a6ce79a-afdb-4e21-9e24-8de6924279bf" (UID: "8a6ce79a-afdb-4e21-9e24-8de6924279bf"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.845611 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-5k5ph"] Mar 12 00:13:51 crc kubenswrapper[4948]: I0312 00:13:51.896281 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/8a6ce79a-afdb-4e21-9e24-8de6924279bf-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 00:13:52 crc kubenswrapper[4948]: I0312 00:13:52.453373 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-zc9zh" event={"ID":"f8cf7ee1-e662-4afd-9e08-112171677c22","Type":"ContainerDied","Data":"759cc547858b754c4e14a3f3cc4a8e48141ee9196217d84770c4d3cfa6e8e8c4"} Mar 12 00:13:52 crc kubenswrapper[4948]: I0312 00:13:52.453433 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-zc9zh" Mar 12 00:13:52 crc kubenswrapper[4948]: I0312 00:13:52.453673 4948 scope.go:117] "RemoveContainer" containerID="326ebf4817bc847d36148437a6dcda1508f716edd2d55a3be608cafdfbfb5f7e" Mar 12 00:13:52 crc kubenswrapper[4948]: I0312 00:13:52.455253 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-5k5ph" event={"ID":"d39a6a83-8fa5-47e4-8047-4b4e713e1a64","Type":"ContainerStarted","Data":"aa10f5af3e99f81c021625e461ea4dd8948ce383ac2c80655ed4e523e9d9af7c"} Mar 12 00:13:52 crc kubenswrapper[4948]: I0312 00:13:52.455356 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-5k5ph" event={"ID":"d39a6a83-8fa5-47e4-8047-4b4e713e1a64","Type":"ContainerStarted","Data":"a27e86e3703a231b924c4efe8586a9002f9fc447830e9c1e0b9292de2802429e"} Mar 12 00:13:52 crc kubenswrapper[4948]: I0312 00:13:52.455389 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/marketplace-operator-79b997595-5k5ph" Mar 12 00:13:52 crc kubenswrapper[4948]: I0312 00:13:52.458386 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/marketplace-operator-79b997595-5k5ph" Mar 12 00:13:52 crc kubenswrapper[4948]: I0312 00:13:52.460773 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-btd4j" Mar 12 00:13:52 crc kubenswrapper[4948]: I0312 00:13:52.460786 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-btd4j" event={"ID":"8a6ce79a-afdb-4e21-9e24-8de6924279bf","Type":"ContainerDied","Data":"1e8fe393396ecc17edae319ad76c0b9fe95e7a850a2db2f6841ecafa4869716d"} Mar 12 00:13:52 crc kubenswrapper[4948]: I0312 00:13:52.462092 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/marketplace-operator-79b997595-zkbhv" Mar 12 00:13:52 crc kubenswrapper[4948]: I0312 00:13:52.462120 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/marketplace-operator-79b997595-zkbhv" event={"ID":"db40ba09-a72f-43d5-8037-e71c96e2bf5b","Type":"ContainerDied","Data":"d150043fc21742d2456851014afeb8e8fe132595929076230128b9b80cd74e7a"} Mar 12 00:13:52 crc kubenswrapper[4948]: I0312 00:13:52.478168 4948 scope.go:117] "RemoveContainer" containerID="53375f01d7cb1d105553e223936295bb8066f47dbacf62b87c6fd8cc659f2202" Mar 12 00:13:52 crc kubenswrapper[4948]: I0312 00:13:52.480059 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-v8h88" Mar 12 00:13:52 crc kubenswrapper[4948]: I0312 00:13:52.482625 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-rflhd" event={"ID":"5184125a-dde1-4db4-9d6b-5b3c45d8d82a","Type":"ContainerDied","Data":"68d6d263e322261e4b32d31022aad0d2424639725173f965e7a064ec610c70e0"} Mar 12 00:13:52 crc kubenswrapper[4948]: I0312 00:13:52.482795 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-rflhd" Mar 12 00:13:52 crc kubenswrapper[4948]: I0312 00:13:52.498692 4948 scope.go:117] "RemoveContainer" containerID="385a4db639f3a271aeab550575db73d3ae75f07ef98cd979694d081c9ec6756e" Mar 12 00:13:52 crc kubenswrapper[4948]: I0312 00:13:52.504209 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/marketplace-operator-79b997595-5k5ph" podStartSLOduration=1.504192586 podStartE2EDuration="1.504192586s" podCreationTimestamp="2026-03-12 00:13:51 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:13:52.498858209 +0000 UTC m=+431.954461947" watchObservedRunningTime="2026-03-12 00:13:52.504192586 +0000 UTC m=+431.959796324" Mar 12 00:13:52 crc kubenswrapper[4948]: I0312 00:13:52.562862 4948 scope.go:117] "RemoveContainer" containerID="b388539342b4cfeef5eebb31026e89abad176dae01f1b209244ea76136260c19" Mar 12 00:13:52 crc kubenswrapper[4948]: I0312 00:13:52.566696 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-zc9zh"] Mar 12 00:13:52 crc kubenswrapper[4948]: I0312 00:13:52.569432 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-zc9zh"] Mar 12 00:13:52 crc kubenswrapper[4948]: I0312 00:13:52.592792 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-zkbhv"] Mar 12 00:13:52 crc kubenswrapper[4948]: I0312 00:13:52.595879 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/marketplace-operator-79b997595-zkbhv"] Mar 12 00:13:52 crc kubenswrapper[4948]: I0312 00:13:52.598870 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-btd4j"] Mar 12 00:13:52 crc kubenswrapper[4948]: I0312 00:13:52.598971 4948 scope.go:117] "RemoveContainer" containerID="f671d445fb2b1163a7bd29f7448a3d3c3956cda4379bfe37ce56af3269fe0c79" Mar 12 00:13:52 crc kubenswrapper[4948]: I0312 00:13:52.602568 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-btd4j"] Mar 12 00:13:52 crc kubenswrapper[4948]: I0312 00:13:52.609807 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-v8h88"] Mar 12 00:13:52 crc kubenswrapper[4948]: I0312 00:13:52.613140 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-v8h88"] Mar 12 00:13:52 crc kubenswrapper[4948]: I0312 00:13:52.622414 4948 scope.go:117] "RemoveContainer" containerID="8157ef88958b3dff32dedd527eed80aa70a58f4f9fa4240a9b952e786b2b6880" Mar 12 00:13:52 crc kubenswrapper[4948]: I0312 00:13:52.623979 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-rflhd"] Mar 12 00:13:52 crc kubenswrapper[4948]: I0312 00:13:52.626327 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-rflhd"] Mar 12 00:13:52 crc kubenswrapper[4948]: I0312 00:13:52.636134 4948 scope.go:117] "RemoveContainer" containerID="6c01edba31d8a4b8c56434839bfda737c5034ad32ad39b244d484ce89accf516" Mar 12 00:13:52 crc kubenswrapper[4948]: I0312 00:13:52.657069 4948 scope.go:117] "RemoveContainer" containerID="bf689a8e9a12667b22bba99be51be88469d4e2cd090bfa6efe2c69a940fe8322" Mar 12 00:13:52 crc kubenswrapper[4948]: I0312 00:13:52.677283 4948 scope.go:117] "RemoveContainer" containerID="34b02d06deb5bf3f40bbae42cf819536fab8af8ed2c6cc5d9120f3ba633efc66" Mar 12 00:13:52 crc kubenswrapper[4948]: I0312 00:13:52.691505 4948 scope.go:117] "RemoveContainer" containerID="f428f193760fb0cfa943b3d75ca100ec3dc34883d4ebe517835fa46ecb35b01e" Mar 12 00:13:53 crc kubenswrapper[4948]: I0312 00:13:53.247341 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-marketplace-qqd6j"] Mar 12 00:13:53 crc kubenswrapper[4948]: E0312 00:13:53.247662 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d83d568-3968-4ea6-b548-511c951defa5" containerName="extract-utilities" Mar 12 00:13:53 crc kubenswrapper[4948]: I0312 00:13:53.247689 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d83d568-3968-4ea6-b548-511c951defa5" containerName="extract-utilities" Mar 12 00:13:53 crc kubenswrapper[4948]: E0312 00:13:53.247709 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5184125a-dde1-4db4-9d6b-5b3c45d8d82a" containerName="registry-server" Mar 12 00:13:53 crc kubenswrapper[4948]: I0312 00:13:53.247722 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="5184125a-dde1-4db4-9d6b-5b3c45d8d82a" containerName="registry-server" Mar 12 00:13:53 crc kubenswrapper[4948]: E0312 00:13:53.247743 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db40ba09-a72f-43d5-8037-e71c96e2bf5b" containerName="marketplace-operator" Mar 12 00:13:53 crc kubenswrapper[4948]: I0312 00:13:53.247755 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="db40ba09-a72f-43d5-8037-e71c96e2bf5b" containerName="marketplace-operator" Mar 12 00:13:53 crc kubenswrapper[4948]: E0312 00:13:53.247777 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5184125a-dde1-4db4-9d6b-5b3c45d8d82a" containerName="extract-content" Mar 12 00:13:53 crc kubenswrapper[4948]: I0312 00:13:53.247790 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="5184125a-dde1-4db4-9d6b-5b3c45d8d82a" containerName="extract-content" Mar 12 00:13:53 crc kubenswrapper[4948]: E0312 00:13:53.247806 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d83d568-3968-4ea6-b548-511c951defa5" containerName="extract-content" Mar 12 00:13:53 crc kubenswrapper[4948]: I0312 00:13:53.247819 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d83d568-3968-4ea6-b548-511c951defa5" containerName="extract-content" Mar 12 00:13:53 crc kubenswrapper[4948]: E0312 00:13:53.247833 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="db40ba09-a72f-43d5-8037-e71c96e2bf5b" containerName="marketplace-operator" Mar 12 00:13:53 crc kubenswrapper[4948]: I0312 00:13:53.247845 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="db40ba09-a72f-43d5-8037-e71c96e2bf5b" containerName="marketplace-operator" Mar 12 00:13:53 crc kubenswrapper[4948]: E0312 00:13:53.247866 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a6ce79a-afdb-4e21-9e24-8de6924279bf" containerName="extract-content" Mar 12 00:13:53 crc kubenswrapper[4948]: I0312 00:13:53.247877 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a6ce79a-afdb-4e21-9e24-8de6924279bf" containerName="extract-content" Mar 12 00:13:53 crc kubenswrapper[4948]: E0312 00:13:53.247892 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5184125a-dde1-4db4-9d6b-5b3c45d8d82a" containerName="extract-utilities" Mar 12 00:13:53 crc kubenswrapper[4948]: I0312 00:13:53.247905 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="5184125a-dde1-4db4-9d6b-5b3c45d8d82a" containerName="extract-utilities" Mar 12 00:13:53 crc kubenswrapper[4948]: E0312 00:13:53.247921 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a6ce79a-afdb-4e21-9e24-8de6924279bf" containerName="extract-utilities" Mar 12 00:13:53 crc kubenswrapper[4948]: I0312 00:13:53.247932 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a6ce79a-afdb-4e21-9e24-8de6924279bf" containerName="extract-utilities" Mar 12 00:13:53 crc kubenswrapper[4948]: E0312 00:13:53.247951 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8cf7ee1-e662-4afd-9e08-112171677c22" containerName="registry-server" Mar 12 00:13:53 crc kubenswrapper[4948]: I0312 00:13:53.247962 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8cf7ee1-e662-4afd-9e08-112171677c22" containerName="registry-server" Mar 12 00:13:53 crc kubenswrapper[4948]: E0312 00:13:53.247976 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8cf7ee1-e662-4afd-9e08-112171677c22" containerName="extract-content" Mar 12 00:13:53 crc kubenswrapper[4948]: I0312 00:13:53.247988 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8cf7ee1-e662-4afd-9e08-112171677c22" containerName="extract-content" Mar 12 00:13:53 crc kubenswrapper[4948]: E0312 00:13:53.248009 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8a6ce79a-afdb-4e21-9e24-8de6924279bf" containerName="registry-server" Mar 12 00:13:53 crc kubenswrapper[4948]: I0312 00:13:53.248021 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="8a6ce79a-afdb-4e21-9e24-8de6924279bf" containerName="registry-server" Mar 12 00:13:53 crc kubenswrapper[4948]: E0312 00:13:53.248039 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5d83d568-3968-4ea6-b548-511c951defa5" containerName="registry-server" Mar 12 00:13:53 crc kubenswrapper[4948]: I0312 00:13:53.248051 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="5d83d568-3968-4ea6-b548-511c951defa5" containerName="registry-server" Mar 12 00:13:53 crc kubenswrapper[4948]: E0312 00:13:53.248070 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f8cf7ee1-e662-4afd-9e08-112171677c22" containerName="extract-utilities" Mar 12 00:13:53 crc kubenswrapper[4948]: I0312 00:13:53.248082 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="f8cf7ee1-e662-4afd-9e08-112171677c22" containerName="extract-utilities" Mar 12 00:13:53 crc kubenswrapper[4948]: I0312 00:13:53.248246 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="5d83d568-3968-4ea6-b548-511c951defa5" containerName="registry-server" Mar 12 00:13:53 crc kubenswrapper[4948]: I0312 00:13:53.248270 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="8a6ce79a-afdb-4e21-9e24-8de6924279bf" containerName="registry-server" Mar 12 00:13:53 crc kubenswrapper[4948]: I0312 00:13:53.248294 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="db40ba09-a72f-43d5-8037-e71c96e2bf5b" containerName="marketplace-operator" Mar 12 00:13:53 crc kubenswrapper[4948]: I0312 00:13:53.248338 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="f8cf7ee1-e662-4afd-9e08-112171677c22" containerName="registry-server" Mar 12 00:13:53 crc kubenswrapper[4948]: I0312 00:13:53.248354 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="db40ba09-a72f-43d5-8037-e71c96e2bf5b" containerName="marketplace-operator" Mar 12 00:13:53 crc kubenswrapper[4948]: I0312 00:13:53.248373 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="5184125a-dde1-4db4-9d6b-5b3c45d8d82a" containerName="registry-server" Mar 12 00:13:53 crc kubenswrapper[4948]: I0312 00:13:53.249625 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qqd6j" Mar 12 00:13:53 crc kubenswrapper[4948]: I0312 00:13:53.251694 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-marketplace-dockercfg-x2ctb" Mar 12 00:13:53 crc kubenswrapper[4948]: I0312 00:13:53.253636 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qqd6j"] Mar 12 00:13:53 crc kubenswrapper[4948]: I0312 00:13:53.319370 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5184125a-dde1-4db4-9d6b-5b3c45d8d82a" path="/var/lib/kubelet/pods/5184125a-dde1-4db4-9d6b-5b3c45d8d82a/volumes" Mar 12 00:13:53 crc kubenswrapper[4948]: I0312 00:13:53.320399 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="5d83d568-3968-4ea6-b548-511c951defa5" path="/var/lib/kubelet/pods/5d83d568-3968-4ea6-b548-511c951defa5/volumes" Mar 12 00:13:53 crc kubenswrapper[4948]: I0312 00:13:53.321223 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="8a6ce79a-afdb-4e21-9e24-8de6924279bf" path="/var/lib/kubelet/pods/8a6ce79a-afdb-4e21-9e24-8de6924279bf/volumes" Mar 12 00:13:53 crc kubenswrapper[4948]: I0312 00:13:53.322712 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="db40ba09-a72f-43d5-8037-e71c96e2bf5b" path="/var/lib/kubelet/pods/db40ba09-a72f-43d5-8037-e71c96e2bf5b/volumes" Mar 12 00:13:53 crc kubenswrapper[4948]: I0312 00:13:53.323339 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f8cf7ee1-e662-4afd-9e08-112171677c22" path="/var/lib/kubelet/pods/f8cf7ee1-e662-4afd-9e08-112171677c22/volumes" Mar 12 00:13:53 crc kubenswrapper[4948]: I0312 00:13:53.323356 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67ac8dbf-c516-4b07-aa10-546e57acdf69-catalog-content\") pod \"redhat-marketplace-qqd6j\" (UID: \"67ac8dbf-c516-4b07-aa10-546e57acdf69\") " pod="openshift-marketplace/redhat-marketplace-qqd6j" Mar 12 00:13:53 crc kubenswrapper[4948]: I0312 00:13:53.323404 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67ac8dbf-c516-4b07-aa10-546e57acdf69-utilities\") pod \"redhat-marketplace-qqd6j\" (UID: \"67ac8dbf-c516-4b07-aa10-546e57acdf69\") " pod="openshift-marketplace/redhat-marketplace-qqd6j" Mar 12 00:13:53 crc kubenswrapper[4948]: I0312 00:13:53.323485 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-s6x58\" (UniqueName: \"kubernetes.io/projected/67ac8dbf-c516-4b07-aa10-546e57acdf69-kube-api-access-s6x58\") pod \"redhat-marketplace-qqd6j\" (UID: \"67ac8dbf-c516-4b07-aa10-546e57acdf69\") " pod="openshift-marketplace/redhat-marketplace-qqd6j" Mar 12 00:13:53 crc kubenswrapper[4948]: I0312 00:13:53.424806 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-s6x58\" (UniqueName: \"kubernetes.io/projected/67ac8dbf-c516-4b07-aa10-546e57acdf69-kube-api-access-s6x58\") pod \"redhat-marketplace-qqd6j\" (UID: \"67ac8dbf-c516-4b07-aa10-546e57acdf69\") " pod="openshift-marketplace/redhat-marketplace-qqd6j" Mar 12 00:13:53 crc kubenswrapper[4948]: I0312 00:13:53.425038 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67ac8dbf-c516-4b07-aa10-546e57acdf69-catalog-content\") pod \"redhat-marketplace-qqd6j\" (UID: \"67ac8dbf-c516-4b07-aa10-546e57acdf69\") " pod="openshift-marketplace/redhat-marketplace-qqd6j" Mar 12 00:13:53 crc kubenswrapper[4948]: I0312 00:13:53.425080 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67ac8dbf-c516-4b07-aa10-546e57acdf69-utilities\") pod \"redhat-marketplace-qqd6j\" (UID: \"67ac8dbf-c516-4b07-aa10-546e57acdf69\") " pod="openshift-marketplace/redhat-marketplace-qqd6j" Mar 12 00:13:53 crc kubenswrapper[4948]: I0312 00:13:53.425511 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67ac8dbf-c516-4b07-aa10-546e57acdf69-catalog-content\") pod \"redhat-marketplace-qqd6j\" (UID: \"67ac8dbf-c516-4b07-aa10-546e57acdf69\") " pod="openshift-marketplace/redhat-marketplace-qqd6j" Mar 12 00:13:53 crc kubenswrapper[4948]: I0312 00:13:53.425741 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67ac8dbf-c516-4b07-aa10-546e57acdf69-utilities\") pod \"redhat-marketplace-qqd6j\" (UID: \"67ac8dbf-c516-4b07-aa10-546e57acdf69\") " pod="openshift-marketplace/redhat-marketplace-qqd6j" Mar 12 00:13:53 crc kubenswrapper[4948]: I0312 00:13:53.441296 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-zpktj"] Mar 12 00:13:53 crc kubenswrapper[4948]: I0312 00:13:53.442971 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zpktj" Mar 12 00:13:53 crc kubenswrapper[4948]: I0312 00:13:53.451112 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"certified-operators-dockercfg-4rs5g" Mar 12 00:13:53 crc kubenswrapper[4948]: I0312 00:13:53.460430 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-s6x58\" (UniqueName: \"kubernetes.io/projected/67ac8dbf-c516-4b07-aa10-546e57acdf69-kube-api-access-s6x58\") pod \"redhat-marketplace-qqd6j\" (UID: \"67ac8dbf-c516-4b07-aa10-546e57acdf69\") " pod="openshift-marketplace/redhat-marketplace-qqd6j" Mar 12 00:13:53 crc kubenswrapper[4948]: I0312 00:13:53.476476 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zpktj"] Mar 12 00:13:53 crc kubenswrapper[4948]: I0312 00:13:53.525701 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3fc280b5-e8c9-4046-bd99-74a0fcbc73ff-utilities\") pod \"certified-operators-zpktj\" (UID: \"3fc280b5-e8c9-4046-bd99-74a0fcbc73ff\") " pod="openshift-marketplace/certified-operators-zpktj" Mar 12 00:13:53 crc kubenswrapper[4948]: I0312 00:13:53.525783 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-28pp8\" (UniqueName: \"kubernetes.io/projected/3fc280b5-e8c9-4046-bd99-74a0fcbc73ff-kube-api-access-28pp8\") pod \"certified-operators-zpktj\" (UID: \"3fc280b5-e8c9-4046-bd99-74a0fcbc73ff\") " pod="openshift-marketplace/certified-operators-zpktj" Mar 12 00:13:53 crc kubenswrapper[4948]: I0312 00:13:53.525876 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3fc280b5-e8c9-4046-bd99-74a0fcbc73ff-catalog-content\") pod \"certified-operators-zpktj\" (UID: \"3fc280b5-e8c9-4046-bd99-74a0fcbc73ff\") " pod="openshift-marketplace/certified-operators-zpktj" Mar 12 00:13:53 crc kubenswrapper[4948]: I0312 00:13:53.576054 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qqd6j" Mar 12 00:13:53 crc kubenswrapper[4948]: I0312 00:13:53.627179 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3fc280b5-e8c9-4046-bd99-74a0fcbc73ff-utilities\") pod \"certified-operators-zpktj\" (UID: \"3fc280b5-e8c9-4046-bd99-74a0fcbc73ff\") " pod="openshift-marketplace/certified-operators-zpktj" Mar 12 00:13:53 crc kubenswrapper[4948]: I0312 00:13:53.627255 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-28pp8\" (UniqueName: \"kubernetes.io/projected/3fc280b5-e8c9-4046-bd99-74a0fcbc73ff-kube-api-access-28pp8\") pod \"certified-operators-zpktj\" (UID: \"3fc280b5-e8c9-4046-bd99-74a0fcbc73ff\") " pod="openshift-marketplace/certified-operators-zpktj" Mar 12 00:13:53 crc kubenswrapper[4948]: I0312 00:13:53.627338 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3fc280b5-e8c9-4046-bd99-74a0fcbc73ff-catalog-content\") pod \"certified-operators-zpktj\" (UID: \"3fc280b5-e8c9-4046-bd99-74a0fcbc73ff\") " pod="openshift-marketplace/certified-operators-zpktj" Mar 12 00:13:53 crc kubenswrapper[4948]: I0312 00:13:53.627629 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/3fc280b5-e8c9-4046-bd99-74a0fcbc73ff-utilities\") pod \"certified-operators-zpktj\" (UID: \"3fc280b5-e8c9-4046-bd99-74a0fcbc73ff\") " pod="openshift-marketplace/certified-operators-zpktj" Mar 12 00:13:53 crc kubenswrapper[4948]: I0312 00:13:53.627737 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/3fc280b5-e8c9-4046-bd99-74a0fcbc73ff-catalog-content\") pod \"certified-operators-zpktj\" (UID: \"3fc280b5-e8c9-4046-bd99-74a0fcbc73ff\") " pod="openshift-marketplace/certified-operators-zpktj" Mar 12 00:13:53 crc kubenswrapper[4948]: I0312 00:13:53.647164 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-28pp8\" (UniqueName: \"kubernetes.io/projected/3fc280b5-e8c9-4046-bd99-74a0fcbc73ff-kube-api-access-28pp8\") pod \"certified-operators-zpktj\" (UID: \"3fc280b5-e8c9-4046-bd99-74a0fcbc73ff\") " pod="openshift-marketplace/certified-operators-zpktj" Mar 12 00:13:53 crc kubenswrapper[4948]: I0312 00:13:53.800853 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-zpktj" Mar 12 00:13:53 crc kubenswrapper[4948]: I0312 00:13:53.814825 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-marketplace-qqd6j"] Mar 12 00:13:53 crc kubenswrapper[4948]: W0312 00:13:53.824685 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod67ac8dbf_c516_4b07_aa10_546e57acdf69.slice/crio-8dc7620762f9f98691bcb63f014c023becf9e59ab861dba28deaef945fd60223 WatchSource:0}: Error finding container 8dc7620762f9f98691bcb63f014c023becf9e59ab861dba28deaef945fd60223: Status 404 returned error can't find the container with id 8dc7620762f9f98691bcb63f014c023becf9e59ab861dba28deaef945fd60223 Mar 12 00:13:53 crc kubenswrapper[4948]: I0312 00:13:53.995238 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-zpktj"] Mar 12 00:13:54 crc kubenswrapper[4948]: W0312 00:13:54.048574 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3fc280b5_e8c9_4046_bd99_74a0fcbc73ff.slice/crio-7b33d3e11e6a5ffa31027af5cac4440bba3c3307ec6520bd1e4e52a737d42d98 WatchSource:0}: Error finding container 7b33d3e11e6a5ffa31027af5cac4440bba3c3307ec6520bd1e4e52a737d42d98: Status 404 returned error can't find the container with id 7b33d3e11e6a5ffa31027af5cac4440bba3c3307ec6520bd1e4e52a737d42d98 Mar 12 00:13:54 crc kubenswrapper[4948]: I0312 00:13:54.509187 4948 generic.go:334] "Generic (PLEG): container finished" podID="67ac8dbf-c516-4b07-aa10-546e57acdf69" containerID="6930ca7436e07739b40d3e73ef2dab180d025594762453c1b8ac6f0bacd45ef0" exitCode=0 Mar 12 00:13:54 crc kubenswrapper[4948]: I0312 00:13:54.509352 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qqd6j" event={"ID":"67ac8dbf-c516-4b07-aa10-546e57acdf69","Type":"ContainerDied","Data":"6930ca7436e07739b40d3e73ef2dab180d025594762453c1b8ac6f0bacd45ef0"} Mar 12 00:13:54 crc kubenswrapper[4948]: I0312 00:13:54.510283 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qqd6j" event={"ID":"67ac8dbf-c516-4b07-aa10-546e57acdf69","Type":"ContainerStarted","Data":"8dc7620762f9f98691bcb63f014c023becf9e59ab861dba28deaef945fd60223"} Mar 12 00:13:54 crc kubenswrapper[4948]: I0312 00:13:54.512177 4948 generic.go:334] "Generic (PLEG): container finished" podID="3fc280b5-e8c9-4046-bd99-74a0fcbc73ff" containerID="3e9e5c8e9f61d8b7a756cd40089a21dae44dc080c34a72031a5b8ec054e0a2eb" exitCode=0 Mar 12 00:13:54 crc kubenswrapper[4948]: I0312 00:13:54.512374 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zpktj" event={"ID":"3fc280b5-e8c9-4046-bd99-74a0fcbc73ff","Type":"ContainerDied","Data":"3e9e5c8e9f61d8b7a756cd40089a21dae44dc080c34a72031a5b8ec054e0a2eb"} Mar 12 00:13:54 crc kubenswrapper[4948]: I0312 00:13:54.512423 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zpktj" event={"ID":"3fc280b5-e8c9-4046-bd99-74a0fcbc73ff","Type":"ContainerStarted","Data":"7b33d3e11e6a5ffa31027af5cac4440bba3c3307ec6520bd1e4e52a737d42d98"} Mar 12 00:13:55 crc kubenswrapper[4948]: I0312 00:13:55.520159 4948 generic.go:334] "Generic (PLEG): container finished" podID="67ac8dbf-c516-4b07-aa10-546e57acdf69" containerID="a31ca9093f6c6f6d5a0728dee03e6c8a89881385ec908b95769d354322c96b6c" exitCode=0 Mar 12 00:13:55 crc kubenswrapper[4948]: I0312 00:13:55.520285 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qqd6j" event={"ID":"67ac8dbf-c516-4b07-aa10-546e57acdf69","Type":"ContainerDied","Data":"a31ca9093f6c6f6d5a0728dee03e6c8a89881385ec908b95769d354322c96b6c"} Mar 12 00:13:55 crc kubenswrapper[4948]: I0312 00:13:55.522814 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zpktj" event={"ID":"3fc280b5-e8c9-4046-bd99-74a0fcbc73ff","Type":"ContainerStarted","Data":"311e187b94a9479eced567b773889a3083a4994dfc985a3fb8bb0bd3d9539fad"} Mar 12 00:13:55 crc kubenswrapper[4948]: E0312 00:13:55.610949 4948 cadvisor_stats_provider.go:516] "Partial failure issuing cadvisor.ContainerInfoV2" err="partial failures: [\"/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3fc280b5_e8c9_4046_bd99_74a0fcbc73ff.slice/crio-311e187b94a9479eced567b773889a3083a4994dfc985a3fb8bb0bd3d9539fad.scope\": RecentStats: unable to find data in memory cache]" Mar 12 00:13:55 crc kubenswrapper[4948]: I0312 00:13:55.834355 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-vmmhs"] Mar 12 00:13:55 crc kubenswrapper[4948]: I0312 00:13:55.835395 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vmmhs" Mar 12 00:13:55 crc kubenswrapper[4948]: I0312 00:13:55.838527 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"redhat-operators-dockercfg-ct8rh" Mar 12 00:13:55 crc kubenswrapper[4948]: I0312 00:13:55.851328 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vmmhs"] Mar 12 00:13:55 crc kubenswrapper[4948]: I0312 00:13:55.858214 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-96fxr\" (UniqueName: \"kubernetes.io/projected/da5b5036-fd96-4ac1-9567-46b309cdf00d-kube-api-access-96fxr\") pod \"redhat-operators-vmmhs\" (UID: \"da5b5036-fd96-4ac1-9567-46b309cdf00d\") " pod="openshift-marketplace/redhat-operators-vmmhs" Mar 12 00:13:55 crc kubenswrapper[4948]: I0312 00:13:55.858292 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da5b5036-fd96-4ac1-9567-46b309cdf00d-catalog-content\") pod \"redhat-operators-vmmhs\" (UID: \"da5b5036-fd96-4ac1-9567-46b309cdf00d\") " pod="openshift-marketplace/redhat-operators-vmmhs" Mar 12 00:13:55 crc kubenswrapper[4948]: I0312 00:13:55.858417 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da5b5036-fd96-4ac1-9567-46b309cdf00d-utilities\") pod \"redhat-operators-vmmhs\" (UID: \"da5b5036-fd96-4ac1-9567-46b309cdf00d\") " pod="openshift-marketplace/redhat-operators-vmmhs" Mar 12 00:13:55 crc kubenswrapper[4948]: I0312 00:13:55.959339 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-96fxr\" (UniqueName: \"kubernetes.io/projected/da5b5036-fd96-4ac1-9567-46b309cdf00d-kube-api-access-96fxr\") pod \"redhat-operators-vmmhs\" (UID: \"da5b5036-fd96-4ac1-9567-46b309cdf00d\") " pod="openshift-marketplace/redhat-operators-vmmhs" Mar 12 00:13:55 crc kubenswrapper[4948]: I0312 00:13:55.959422 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da5b5036-fd96-4ac1-9567-46b309cdf00d-catalog-content\") pod \"redhat-operators-vmmhs\" (UID: \"da5b5036-fd96-4ac1-9567-46b309cdf00d\") " pod="openshift-marketplace/redhat-operators-vmmhs" Mar 12 00:13:55 crc kubenswrapper[4948]: I0312 00:13:55.959638 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da5b5036-fd96-4ac1-9567-46b309cdf00d-utilities\") pod \"redhat-operators-vmmhs\" (UID: \"da5b5036-fd96-4ac1-9567-46b309cdf00d\") " pod="openshift-marketplace/redhat-operators-vmmhs" Mar 12 00:13:55 crc kubenswrapper[4948]: I0312 00:13:55.960196 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/da5b5036-fd96-4ac1-9567-46b309cdf00d-utilities\") pod \"redhat-operators-vmmhs\" (UID: \"da5b5036-fd96-4ac1-9567-46b309cdf00d\") " pod="openshift-marketplace/redhat-operators-vmmhs" Mar 12 00:13:55 crc kubenswrapper[4948]: I0312 00:13:55.960532 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/da5b5036-fd96-4ac1-9567-46b309cdf00d-catalog-content\") pod \"redhat-operators-vmmhs\" (UID: \"da5b5036-fd96-4ac1-9567-46b309cdf00d\") " pod="openshift-marketplace/redhat-operators-vmmhs" Mar 12 00:13:55 crc kubenswrapper[4948]: I0312 00:13:55.990882 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-96fxr\" (UniqueName: \"kubernetes.io/projected/da5b5036-fd96-4ac1-9567-46b309cdf00d-kube-api-access-96fxr\") pod \"redhat-operators-vmmhs\" (UID: \"da5b5036-fd96-4ac1-9567-46b309cdf00d\") " pod="openshift-marketplace/redhat-operators-vmmhs" Mar 12 00:13:56 crc kubenswrapper[4948]: I0312 00:13:56.042742 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-dc79v"] Mar 12 00:13:56 crc kubenswrapper[4948]: I0312 00:13:56.044149 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dc79v" Mar 12 00:13:56 crc kubenswrapper[4948]: I0312 00:13:56.046810 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"community-operators-dockercfg-dmngl" Mar 12 00:13:56 crc kubenswrapper[4948]: I0312 00:13:56.054748 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dc79v"] Mar 12 00:13:56 crc kubenswrapper[4948]: I0312 00:13:56.091932 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6658e419-c016-4b00-bb2c-24ffdbc65283-catalog-content\") pod \"community-operators-dc79v\" (UID: \"6658e419-c016-4b00-bb2c-24ffdbc65283\") " pod="openshift-marketplace/community-operators-dc79v" Mar 12 00:13:56 crc kubenswrapper[4948]: I0312 00:13:56.092020 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6658e419-c016-4b00-bb2c-24ffdbc65283-utilities\") pod \"community-operators-dc79v\" (UID: \"6658e419-c016-4b00-bb2c-24ffdbc65283\") " pod="openshift-marketplace/community-operators-dc79v" Mar 12 00:13:56 crc kubenswrapper[4948]: I0312 00:13:56.092048 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jzpvq\" (UniqueName: \"kubernetes.io/projected/6658e419-c016-4b00-bb2c-24ffdbc65283-kube-api-access-jzpvq\") pod \"community-operators-dc79v\" (UID: \"6658e419-c016-4b00-bb2c-24ffdbc65283\") " pod="openshift-marketplace/community-operators-dc79v" Mar 12 00:13:56 crc kubenswrapper[4948]: I0312 00:13:56.170573 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-vmmhs" Mar 12 00:13:56 crc kubenswrapper[4948]: I0312 00:13:56.193177 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6658e419-c016-4b00-bb2c-24ffdbc65283-utilities\") pod \"community-operators-dc79v\" (UID: \"6658e419-c016-4b00-bb2c-24ffdbc65283\") " pod="openshift-marketplace/community-operators-dc79v" Mar 12 00:13:56 crc kubenswrapper[4948]: I0312 00:13:56.193368 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jzpvq\" (UniqueName: \"kubernetes.io/projected/6658e419-c016-4b00-bb2c-24ffdbc65283-kube-api-access-jzpvq\") pod \"community-operators-dc79v\" (UID: \"6658e419-c016-4b00-bb2c-24ffdbc65283\") " pod="openshift-marketplace/community-operators-dc79v" Mar 12 00:13:56 crc kubenswrapper[4948]: I0312 00:13:56.193431 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6658e419-c016-4b00-bb2c-24ffdbc65283-catalog-content\") pod \"community-operators-dc79v\" (UID: \"6658e419-c016-4b00-bb2c-24ffdbc65283\") " pod="openshift-marketplace/community-operators-dc79v" Mar 12 00:13:56 crc kubenswrapper[4948]: I0312 00:13:56.193667 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/6658e419-c016-4b00-bb2c-24ffdbc65283-utilities\") pod \"community-operators-dc79v\" (UID: \"6658e419-c016-4b00-bb2c-24ffdbc65283\") " pod="openshift-marketplace/community-operators-dc79v" Mar 12 00:13:56 crc kubenswrapper[4948]: I0312 00:13:56.193831 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/6658e419-c016-4b00-bb2c-24ffdbc65283-catalog-content\") pod \"community-operators-dc79v\" (UID: \"6658e419-c016-4b00-bb2c-24ffdbc65283\") " pod="openshift-marketplace/community-operators-dc79v" Mar 12 00:13:56 crc kubenswrapper[4948]: I0312 00:13:56.226243 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jzpvq\" (UniqueName: \"kubernetes.io/projected/6658e419-c016-4b00-bb2c-24ffdbc65283-kube-api-access-jzpvq\") pod \"community-operators-dc79v\" (UID: \"6658e419-c016-4b00-bb2c-24ffdbc65283\") " pod="openshift-marketplace/community-operators-dc79v" Mar 12 00:13:56 crc kubenswrapper[4948]: I0312 00:13:56.368794 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-vmmhs"] Mar 12 00:13:56 crc kubenswrapper[4948]: I0312 00:13:56.435634 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-dc79v" Mar 12 00:13:56 crc kubenswrapper[4948]: I0312 00:13:56.534187 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qqd6j" event={"ID":"67ac8dbf-c516-4b07-aa10-546e57acdf69","Type":"ContainerStarted","Data":"304db3562a72e39f59cfa721deef292badcb7adfdbf6b21f029f64c6001a1640"} Mar 12 00:13:56 crc kubenswrapper[4948]: I0312 00:13:56.537671 4948 generic.go:334] "Generic (PLEG): container finished" podID="3fc280b5-e8c9-4046-bd99-74a0fcbc73ff" containerID="311e187b94a9479eced567b773889a3083a4994dfc985a3fb8bb0bd3d9539fad" exitCode=0 Mar 12 00:13:56 crc kubenswrapper[4948]: I0312 00:13:56.537754 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zpktj" event={"ID":"3fc280b5-e8c9-4046-bd99-74a0fcbc73ff","Type":"ContainerDied","Data":"311e187b94a9479eced567b773889a3083a4994dfc985a3fb8bb0bd3d9539fad"} Mar 12 00:13:56 crc kubenswrapper[4948]: I0312 00:13:56.541101 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vmmhs" event={"ID":"da5b5036-fd96-4ac1-9567-46b309cdf00d","Type":"ContainerStarted","Data":"1c1ee60b2421afc8e4419bd81a1493b4454f059f7b3cf36ffe3ae314b2b9354e"} Mar 12 00:13:56 crc kubenswrapper[4948]: I0312 00:13:56.541133 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vmmhs" event={"ID":"da5b5036-fd96-4ac1-9567-46b309cdf00d","Type":"ContainerStarted","Data":"2a9551394a47bdb87004ba64c5b070a5ffe4f69dac65facb4472368819b444c5"} Mar 12 00:13:56 crc kubenswrapper[4948]: I0312 00:13:56.555004 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-marketplace-qqd6j" podStartSLOduration=2.117736993 podStartE2EDuration="3.554986996s" podCreationTimestamp="2026-03-12 00:13:53 +0000 UTC" firstStartedPulling="2026-03-12 00:13:54.510872986 +0000 UTC m=+433.966476724" lastFinishedPulling="2026-03-12 00:13:55.948122989 +0000 UTC m=+435.403726727" observedRunningTime="2026-03-12 00:13:56.55044139 +0000 UTC m=+436.006045138" watchObservedRunningTime="2026-03-12 00:13:56.554986996 +0000 UTC m=+436.010590744" Mar 12 00:13:56 crc kubenswrapper[4948]: I0312 00:13:56.645832 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-dc79v"] Mar 12 00:13:57 crc kubenswrapper[4948]: I0312 00:13:57.547408 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-zpktj" event={"ID":"3fc280b5-e8c9-4046-bd99-74a0fcbc73ff","Type":"ContainerStarted","Data":"5babdaece9af57fc74d94509b919841525091add0b7b67b87622f29d054063fe"} Mar 12 00:13:57 crc kubenswrapper[4948]: I0312 00:13:57.548766 4948 generic.go:334] "Generic (PLEG): container finished" podID="da5b5036-fd96-4ac1-9567-46b309cdf00d" containerID="1c1ee60b2421afc8e4419bd81a1493b4454f059f7b3cf36ffe3ae314b2b9354e" exitCode=0 Mar 12 00:13:57 crc kubenswrapper[4948]: I0312 00:13:57.548831 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vmmhs" event={"ID":"da5b5036-fd96-4ac1-9567-46b309cdf00d","Type":"ContainerDied","Data":"1c1ee60b2421afc8e4419bd81a1493b4454f059f7b3cf36ffe3ae314b2b9354e"} Mar 12 00:13:57 crc kubenswrapper[4948]: I0312 00:13:57.549918 4948 generic.go:334] "Generic (PLEG): container finished" podID="6658e419-c016-4b00-bb2c-24ffdbc65283" containerID="718ab912bef15458cab9d09a2f38907d2adf6eb0ea883ff3bb2102fc43ab1cca" exitCode=0 Mar 12 00:13:57 crc kubenswrapper[4948]: I0312 00:13:57.550596 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dc79v" event={"ID":"6658e419-c016-4b00-bb2c-24ffdbc65283","Type":"ContainerDied","Data":"718ab912bef15458cab9d09a2f38907d2adf6eb0ea883ff3bb2102fc43ab1cca"} Mar 12 00:13:57 crc kubenswrapper[4948]: I0312 00:13:57.550616 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dc79v" event={"ID":"6658e419-c016-4b00-bb2c-24ffdbc65283","Type":"ContainerStarted","Data":"6a90c6f005eda65835d37a0a78f2b01de5a99ddacc476468e59fc2d45efb33fd"} Mar 12 00:13:57 crc kubenswrapper[4948]: I0312 00:13:57.569704 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-zpktj" podStartSLOduration=2.14887593 podStartE2EDuration="4.569681488s" podCreationTimestamp="2026-03-12 00:13:53 +0000 UTC" firstStartedPulling="2026-03-12 00:13:54.513953344 +0000 UTC m=+433.969557122" lastFinishedPulling="2026-03-12 00:13:56.934758922 +0000 UTC m=+436.390362680" observedRunningTime="2026-03-12 00:13:57.564564326 +0000 UTC m=+437.020168064" watchObservedRunningTime="2026-03-12 00:13:57.569681488 +0000 UTC m=+437.025285226" Mar 12 00:13:58 crc kubenswrapper[4948]: I0312 00:13:58.556606 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vmmhs" event={"ID":"da5b5036-fd96-4ac1-9567-46b309cdf00d","Type":"ContainerStarted","Data":"1ea65ab246104675ed40736ac5ecdbb9f8375a89bfa12f85e35d56717404be75"} Mar 12 00:13:58 crc kubenswrapper[4948]: I0312 00:13:58.559453 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dc79v" event={"ID":"6658e419-c016-4b00-bb2c-24ffdbc65283","Type":"ContainerStarted","Data":"546c6201efce864282f5d3ac0c0d1e0f4bcf365b5139506148b3944d281552e0"} Mar 12 00:13:59 crc kubenswrapper[4948]: I0312 00:13:59.579126 4948 generic.go:334] "Generic (PLEG): container finished" podID="da5b5036-fd96-4ac1-9567-46b309cdf00d" containerID="1ea65ab246104675ed40736ac5ecdbb9f8375a89bfa12f85e35d56717404be75" exitCode=0 Mar 12 00:13:59 crc kubenswrapper[4948]: I0312 00:13:59.579231 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vmmhs" event={"ID":"da5b5036-fd96-4ac1-9567-46b309cdf00d","Type":"ContainerDied","Data":"1ea65ab246104675ed40736ac5ecdbb9f8375a89bfa12f85e35d56717404be75"} Mar 12 00:13:59 crc kubenswrapper[4948]: I0312 00:13:59.589142 4948 generic.go:334] "Generic (PLEG): container finished" podID="6658e419-c016-4b00-bb2c-24ffdbc65283" containerID="546c6201efce864282f5d3ac0c0d1e0f4bcf365b5139506148b3944d281552e0" exitCode=0 Mar 12 00:13:59 crc kubenswrapper[4948]: I0312 00:13:59.589268 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dc79v" event={"ID":"6658e419-c016-4b00-bb2c-24ffdbc65283","Type":"ContainerDied","Data":"546c6201efce864282f5d3ac0c0d1e0f4bcf365b5139506148b3944d281552e0"} Mar 12 00:14:00 crc kubenswrapper[4948]: I0312 00:14:00.147455 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29554574-5h7vf"] Mar 12 00:14:00 crc kubenswrapper[4948]: I0312 00:14:00.149063 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29554574-5h7vf" Mar 12 00:14:00 crc kubenswrapper[4948]: I0312 00:14:00.151744 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-44g5m" Mar 12 00:14:00 crc kubenswrapper[4948]: I0312 00:14:00.153111 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 00:14:00 crc kubenswrapper[4948]: I0312 00:14:00.153385 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 00:14:00 crc kubenswrapper[4948]: I0312 00:14:00.168468 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29554574-5h7vf"] Mar 12 00:14:00 crc kubenswrapper[4948]: I0312 00:14:00.347544 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-qbmdk\" (UniqueName: \"kubernetes.io/projected/f9a50fc1-6ae2-4bb0-91f5-2e51b77d1b36-kube-api-access-qbmdk\") pod \"auto-csr-approver-29554574-5h7vf\" (UID: \"f9a50fc1-6ae2-4bb0-91f5-2e51b77d1b36\") " pod="openshift-infra/auto-csr-approver-29554574-5h7vf" Mar 12 00:14:00 crc kubenswrapper[4948]: I0312 00:14:00.449525 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-qbmdk\" (UniqueName: \"kubernetes.io/projected/f9a50fc1-6ae2-4bb0-91f5-2e51b77d1b36-kube-api-access-qbmdk\") pod \"auto-csr-approver-29554574-5h7vf\" (UID: \"f9a50fc1-6ae2-4bb0-91f5-2e51b77d1b36\") " pod="openshift-infra/auto-csr-approver-29554574-5h7vf" Mar 12 00:14:00 crc kubenswrapper[4948]: I0312 00:14:00.484834 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-qbmdk\" (UniqueName: \"kubernetes.io/projected/f9a50fc1-6ae2-4bb0-91f5-2e51b77d1b36-kube-api-access-qbmdk\") pod \"auto-csr-approver-29554574-5h7vf\" (UID: \"f9a50fc1-6ae2-4bb0-91f5-2e51b77d1b36\") " pod="openshift-infra/auto-csr-approver-29554574-5h7vf" Mar 12 00:14:00 crc kubenswrapper[4948]: I0312 00:14:00.517059 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29554574-5h7vf" Mar 12 00:14:00 crc kubenswrapper[4948]: I0312 00:14:00.611047 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-dc79v" event={"ID":"6658e419-c016-4b00-bb2c-24ffdbc65283","Type":"ContainerStarted","Data":"5d0825c8556127b11737a77851b75c8674da8514b11b61b879d7fbd7b4d8fdbf"} Mar 12 00:14:00 crc kubenswrapper[4948]: I0312 00:14:00.618087 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-vmmhs" event={"ID":"da5b5036-fd96-4ac1-9567-46b309cdf00d","Type":"ContainerStarted","Data":"6a479656dd3daa4561ec3061bc7aae630f22c4c37bf220672c7bfa2070608f89"} Mar 12 00:14:00 crc kubenswrapper[4948]: I0312 00:14:00.643190 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-dc79v" podStartSLOduration=2.157966369 podStartE2EDuration="4.643172575s" podCreationTimestamp="2026-03-12 00:13:56 +0000 UTC" firstStartedPulling="2026-03-12 00:13:57.551252275 +0000 UTC m=+437.006856023" lastFinishedPulling="2026-03-12 00:14:00.036458451 +0000 UTC m=+439.492062229" observedRunningTime="2026-03-12 00:14:00.641648596 +0000 UTC m=+440.097252354" watchObservedRunningTime="2026-03-12 00:14:00.643172575 +0000 UTC m=+440.098776313" Mar 12 00:14:00 crc kubenswrapper[4948]: I0312 00:14:00.668365 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-vmmhs" podStartSLOduration=3.164267801 podStartE2EDuration="5.668348799s" podCreationTimestamp="2026-03-12 00:13:55 +0000 UTC" firstStartedPulling="2026-03-12 00:13:57.550668291 +0000 UTC m=+437.006272029" lastFinishedPulling="2026-03-12 00:14:00.054749259 +0000 UTC m=+439.510353027" observedRunningTime="2026-03-12 00:14:00.666723716 +0000 UTC m=+440.122327454" watchObservedRunningTime="2026-03-12 00:14:00.668348799 +0000 UTC m=+440.123952537" Mar 12 00:14:01 crc kubenswrapper[4948]: I0312 00:14:01.006033 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29554574-5h7vf"] Mar 12 00:14:01 crc kubenswrapper[4948]: W0312 00:14:01.016195 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-podf9a50fc1_6ae2_4bb0_91f5_2e51b77d1b36.slice/crio-d506261e0008a522d594547998f45e0beb87a55a7cca562e4306bc4f8f304040 WatchSource:0}: Error finding container d506261e0008a522d594547998f45e0beb87a55a7cca562e4306bc4f8f304040: Status 404 returned error can't find the container with id d506261e0008a522d594547998f45e0beb87a55a7cca562e4306bc4f8f304040 Mar 12 00:14:01 crc kubenswrapper[4948]: I0312 00:14:01.634285 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29554574-5h7vf" event={"ID":"f9a50fc1-6ae2-4bb0-91f5-2e51b77d1b36","Type":"ContainerStarted","Data":"d506261e0008a522d594547998f45e0beb87a55a7cca562e4306bc4f8f304040"} Mar 12 00:14:03 crc kubenswrapper[4948]: I0312 00:14:03.577080 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-marketplace-qqd6j" Mar 12 00:14:03 crc kubenswrapper[4948]: I0312 00:14:03.577467 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-marketplace-qqd6j" Mar 12 00:14:03 crc kubenswrapper[4948]: I0312 00:14:03.653178 4948 generic.go:334] "Generic (PLEG): container finished" podID="f9a50fc1-6ae2-4bb0-91f5-2e51b77d1b36" containerID="7a35745a45555a1ae1f5c1b4600afa7ae8e2ab7e7001b750cee96f3375eb7f42" exitCode=0 Mar 12 00:14:03 crc kubenswrapper[4948]: I0312 00:14:03.653239 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29554574-5h7vf" event={"ID":"f9a50fc1-6ae2-4bb0-91f5-2e51b77d1b36","Type":"ContainerDied","Data":"7a35745a45555a1ae1f5c1b4600afa7ae8e2ab7e7001b750cee96f3375eb7f42"} Mar 12 00:14:03 crc kubenswrapper[4948]: I0312 00:14:03.657506 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-marketplace-qqd6j" Mar 12 00:14:03 crc kubenswrapper[4948]: I0312 00:14:03.723675 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-marketplace-qqd6j" Mar 12 00:14:03 crc kubenswrapper[4948]: I0312 00:14:03.801697 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-zpktj" Mar 12 00:14:03 crc kubenswrapper[4948]: I0312 00:14:03.801777 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-zpktj" Mar 12 00:14:03 crc kubenswrapper[4948]: I0312 00:14:03.859017 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-zpktj" Mar 12 00:14:04 crc kubenswrapper[4948]: I0312 00:14:04.712287 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-zpktj" Mar 12 00:14:04 crc kubenswrapper[4948]: I0312 00:14:04.943172 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29554574-5h7vf" Mar 12 00:14:05 crc kubenswrapper[4948]: I0312 00:14:05.107956 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-qbmdk\" (UniqueName: \"kubernetes.io/projected/f9a50fc1-6ae2-4bb0-91f5-2e51b77d1b36-kube-api-access-qbmdk\") pod \"f9a50fc1-6ae2-4bb0-91f5-2e51b77d1b36\" (UID: \"f9a50fc1-6ae2-4bb0-91f5-2e51b77d1b36\") " Mar 12 00:14:05 crc kubenswrapper[4948]: I0312 00:14:05.114493 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/f9a50fc1-6ae2-4bb0-91f5-2e51b77d1b36-kube-api-access-qbmdk" (OuterVolumeSpecName: "kube-api-access-qbmdk") pod "f9a50fc1-6ae2-4bb0-91f5-2e51b77d1b36" (UID: "f9a50fc1-6ae2-4bb0-91f5-2e51b77d1b36"). InnerVolumeSpecName "kube-api-access-qbmdk". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:14:05 crc kubenswrapper[4948]: I0312 00:14:05.210034 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-qbmdk\" (UniqueName: \"kubernetes.io/projected/f9a50fc1-6ae2-4bb0-91f5-2e51b77d1b36-kube-api-access-qbmdk\") on node \"crc\" DevicePath \"\"" Mar 12 00:14:05 crc kubenswrapper[4948]: I0312 00:14:05.671143 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29554574-5h7vf" event={"ID":"f9a50fc1-6ae2-4bb0-91f5-2e51b77d1b36","Type":"ContainerDied","Data":"d506261e0008a522d594547998f45e0beb87a55a7cca562e4306bc4f8f304040"} Mar 12 00:14:05 crc kubenswrapper[4948]: I0312 00:14:05.671646 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="d506261e0008a522d594547998f45e0beb87a55a7cca562e4306bc4f8f304040" Mar 12 00:14:05 crc kubenswrapper[4948]: I0312 00:14:05.671156 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29554574-5h7vf" Mar 12 00:14:06 crc kubenswrapper[4948]: I0312 00:14:06.171147 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-vmmhs" Mar 12 00:14:06 crc kubenswrapper[4948]: I0312 00:14:06.171510 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-vmmhs" Mar 12 00:14:06 crc kubenswrapper[4948]: I0312 00:14:06.436509 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-dc79v" Mar 12 00:14:06 crc kubenswrapper[4948]: I0312 00:14:06.436559 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-dc79v" Mar 12 00:14:06 crc kubenswrapper[4948]: I0312 00:14:06.484078 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-dc79v" Mar 12 00:14:06 crc kubenswrapper[4948]: I0312 00:14:06.731579 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-dc79v" Mar 12 00:14:07 crc kubenswrapper[4948]: I0312 00:14:07.226853 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-vmmhs" podUID="da5b5036-fd96-4ac1-9567-46b309cdf00d" containerName="registry-server" probeResult="failure" output=< Mar 12 00:14:07 crc kubenswrapper[4948]: timeout: failed to connect service ":50051" within 1s Mar 12 00:14:07 crc kubenswrapper[4948]: > Mar 12 00:14:07 crc kubenswrapper[4948]: I0312 00:14:07.979370 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-wb4b8"] Mar 12 00:14:07 crc kubenswrapper[4948]: E0312 00:14:07.979569 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="f9a50fc1-6ae2-4bb0-91f5-2e51b77d1b36" containerName="oc" Mar 12 00:14:07 crc kubenswrapper[4948]: I0312 00:14:07.979580 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="f9a50fc1-6ae2-4bb0-91f5-2e51b77d1b36" containerName="oc" Mar 12 00:14:07 crc kubenswrapper[4948]: I0312 00:14:07.979673 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="f9a50fc1-6ae2-4bb0-91f5-2e51b77d1b36" containerName="oc" Mar 12 00:14:07 crc kubenswrapper[4948]: I0312 00:14:07.980048 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-wb4b8" Mar 12 00:14:07 crc kubenswrapper[4948]: I0312 00:14:07.998188 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-wb4b8"] Mar 12 00:14:08 crc kubenswrapper[4948]: I0312 00:14:08.145962 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/567e59f8-2c9e-4332-8eae-ef3e44f5a851-ca-trust-extracted\") pod \"image-registry-66df7c8f76-wb4b8\" (UID: \"567e59f8-2c9e-4332-8eae-ef3e44f5a851\") " pod="openshift-image-registry/image-registry-66df7c8f76-wb4b8" Mar 12 00:14:08 crc kubenswrapper[4948]: I0312 00:14:08.146020 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-wb4b8\" (UID: \"567e59f8-2c9e-4332-8eae-ef3e44f5a851\") " pod="openshift-image-registry/image-registry-66df7c8f76-wb4b8" Mar 12 00:14:08 crc kubenswrapper[4948]: I0312 00:14:08.146046 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/567e59f8-2c9e-4332-8eae-ef3e44f5a851-installation-pull-secrets\") pod \"image-registry-66df7c8f76-wb4b8\" (UID: \"567e59f8-2c9e-4332-8eae-ef3e44f5a851\") " pod="openshift-image-registry/image-registry-66df7c8f76-wb4b8" Mar 12 00:14:08 crc kubenswrapper[4948]: I0312 00:14:08.146256 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/567e59f8-2c9e-4332-8eae-ef3e44f5a851-bound-sa-token\") pod \"image-registry-66df7c8f76-wb4b8\" (UID: \"567e59f8-2c9e-4332-8eae-ef3e44f5a851\") " pod="openshift-image-registry/image-registry-66df7c8f76-wb4b8" Mar 12 00:14:08 crc kubenswrapper[4948]: I0312 00:14:08.146323 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/567e59f8-2c9e-4332-8eae-ef3e44f5a851-registry-certificates\") pod \"image-registry-66df7c8f76-wb4b8\" (UID: \"567e59f8-2c9e-4332-8eae-ef3e44f5a851\") " pod="openshift-image-registry/image-registry-66df7c8f76-wb4b8" Mar 12 00:14:08 crc kubenswrapper[4948]: I0312 00:14:08.146431 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-lb887\" (UniqueName: \"kubernetes.io/projected/567e59f8-2c9e-4332-8eae-ef3e44f5a851-kube-api-access-lb887\") pod \"image-registry-66df7c8f76-wb4b8\" (UID: \"567e59f8-2c9e-4332-8eae-ef3e44f5a851\") " pod="openshift-image-registry/image-registry-66df7c8f76-wb4b8" Mar 12 00:14:08 crc kubenswrapper[4948]: I0312 00:14:08.146477 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/567e59f8-2c9e-4332-8eae-ef3e44f5a851-registry-tls\") pod \"image-registry-66df7c8f76-wb4b8\" (UID: \"567e59f8-2c9e-4332-8eae-ef3e44f5a851\") " pod="openshift-image-registry/image-registry-66df7c8f76-wb4b8" Mar 12 00:14:08 crc kubenswrapper[4948]: I0312 00:14:08.146550 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/567e59f8-2c9e-4332-8eae-ef3e44f5a851-trusted-ca\") pod \"image-registry-66df7c8f76-wb4b8\" (UID: \"567e59f8-2c9e-4332-8eae-ef3e44f5a851\") " pod="openshift-image-registry/image-registry-66df7c8f76-wb4b8" Mar 12 00:14:08 crc kubenswrapper[4948]: I0312 00:14:08.200154 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"image-registry-66df7c8f76-wb4b8\" (UID: \"567e59f8-2c9e-4332-8eae-ef3e44f5a851\") " pod="openshift-image-registry/image-registry-66df7c8f76-wb4b8" Mar 12 00:14:08 crc kubenswrapper[4948]: I0312 00:14:08.247623 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/567e59f8-2c9e-4332-8eae-ef3e44f5a851-trusted-ca\") pod \"image-registry-66df7c8f76-wb4b8\" (UID: \"567e59f8-2c9e-4332-8eae-ef3e44f5a851\") " pod="openshift-image-registry/image-registry-66df7c8f76-wb4b8" Mar 12 00:14:08 crc kubenswrapper[4948]: I0312 00:14:08.247690 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/567e59f8-2c9e-4332-8eae-ef3e44f5a851-ca-trust-extracted\") pod \"image-registry-66df7c8f76-wb4b8\" (UID: \"567e59f8-2c9e-4332-8eae-ef3e44f5a851\") " pod="openshift-image-registry/image-registry-66df7c8f76-wb4b8" Mar 12 00:14:08 crc kubenswrapper[4948]: I0312 00:14:08.247742 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/567e59f8-2c9e-4332-8eae-ef3e44f5a851-installation-pull-secrets\") pod \"image-registry-66df7c8f76-wb4b8\" (UID: \"567e59f8-2c9e-4332-8eae-ef3e44f5a851\") " pod="openshift-image-registry/image-registry-66df7c8f76-wb4b8" Mar 12 00:14:08 crc kubenswrapper[4948]: I0312 00:14:08.247789 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/567e59f8-2c9e-4332-8eae-ef3e44f5a851-bound-sa-token\") pod \"image-registry-66df7c8f76-wb4b8\" (UID: \"567e59f8-2c9e-4332-8eae-ef3e44f5a851\") " pod="openshift-image-registry/image-registry-66df7c8f76-wb4b8" Mar 12 00:14:08 crc kubenswrapper[4948]: I0312 00:14:08.247811 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/567e59f8-2c9e-4332-8eae-ef3e44f5a851-registry-certificates\") pod \"image-registry-66df7c8f76-wb4b8\" (UID: \"567e59f8-2c9e-4332-8eae-ef3e44f5a851\") " pod="openshift-image-registry/image-registry-66df7c8f76-wb4b8" Mar 12 00:14:08 crc kubenswrapper[4948]: I0312 00:14:08.247862 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-lb887\" (UniqueName: \"kubernetes.io/projected/567e59f8-2c9e-4332-8eae-ef3e44f5a851-kube-api-access-lb887\") pod \"image-registry-66df7c8f76-wb4b8\" (UID: \"567e59f8-2c9e-4332-8eae-ef3e44f5a851\") " pod="openshift-image-registry/image-registry-66df7c8f76-wb4b8" Mar 12 00:14:08 crc kubenswrapper[4948]: I0312 00:14:08.247894 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/567e59f8-2c9e-4332-8eae-ef3e44f5a851-registry-tls\") pod \"image-registry-66df7c8f76-wb4b8\" (UID: \"567e59f8-2c9e-4332-8eae-ef3e44f5a851\") " pod="openshift-image-registry/image-registry-66df7c8f76-wb4b8" Mar 12 00:14:08 crc kubenswrapper[4948]: I0312 00:14:08.248565 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/567e59f8-2c9e-4332-8eae-ef3e44f5a851-ca-trust-extracted\") pod \"image-registry-66df7c8f76-wb4b8\" (UID: \"567e59f8-2c9e-4332-8eae-ef3e44f5a851\") " pod="openshift-image-registry/image-registry-66df7c8f76-wb4b8" Mar 12 00:14:08 crc kubenswrapper[4948]: I0312 00:14:08.249520 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/567e59f8-2c9e-4332-8eae-ef3e44f5a851-trusted-ca\") pod \"image-registry-66df7c8f76-wb4b8\" (UID: \"567e59f8-2c9e-4332-8eae-ef3e44f5a851\") " pod="openshift-image-registry/image-registry-66df7c8f76-wb4b8" Mar 12 00:14:08 crc kubenswrapper[4948]: I0312 00:14:08.249577 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/567e59f8-2c9e-4332-8eae-ef3e44f5a851-registry-certificates\") pod \"image-registry-66df7c8f76-wb4b8\" (UID: \"567e59f8-2c9e-4332-8eae-ef3e44f5a851\") " pod="openshift-image-registry/image-registry-66df7c8f76-wb4b8" Mar 12 00:14:08 crc kubenswrapper[4948]: I0312 00:14:08.254062 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/567e59f8-2c9e-4332-8eae-ef3e44f5a851-installation-pull-secrets\") pod \"image-registry-66df7c8f76-wb4b8\" (UID: \"567e59f8-2c9e-4332-8eae-ef3e44f5a851\") " pod="openshift-image-registry/image-registry-66df7c8f76-wb4b8" Mar 12 00:14:08 crc kubenswrapper[4948]: I0312 00:14:08.264052 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/567e59f8-2c9e-4332-8eae-ef3e44f5a851-bound-sa-token\") pod \"image-registry-66df7c8f76-wb4b8\" (UID: \"567e59f8-2c9e-4332-8eae-ef3e44f5a851\") " pod="openshift-image-registry/image-registry-66df7c8f76-wb4b8" Mar 12 00:14:08 crc kubenswrapper[4948]: I0312 00:14:08.270059 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/567e59f8-2c9e-4332-8eae-ef3e44f5a851-registry-tls\") pod \"image-registry-66df7c8f76-wb4b8\" (UID: \"567e59f8-2c9e-4332-8eae-ef3e44f5a851\") " pod="openshift-image-registry/image-registry-66df7c8f76-wb4b8" Mar 12 00:14:08 crc kubenswrapper[4948]: I0312 00:14:08.275007 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-lb887\" (UniqueName: \"kubernetes.io/projected/567e59f8-2c9e-4332-8eae-ef3e44f5a851-kube-api-access-lb887\") pod \"image-registry-66df7c8f76-wb4b8\" (UID: \"567e59f8-2c9e-4332-8eae-ef3e44f5a851\") " pod="openshift-image-registry/image-registry-66df7c8f76-wb4b8" Mar 12 00:14:08 crc kubenswrapper[4948]: I0312 00:14:08.292463 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-66df7c8f76-wb4b8" Mar 12 00:14:08 crc kubenswrapper[4948]: I0312 00:14:08.950289 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-image-registry/image-registry-66df7c8f76-wb4b8"] Mar 12 00:14:08 crc kubenswrapper[4948]: W0312 00:14:08.966831 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod567e59f8_2c9e_4332_8eae_ef3e44f5a851.slice/crio-6088259c08ed03828378a8ae62f10dfe8b1ea8589685f497766b40b417f9789d WatchSource:0}: Error finding container 6088259c08ed03828378a8ae62f10dfe8b1ea8589685f497766b40b417f9789d: Status 404 returned error can't find the container with id 6088259c08ed03828378a8ae62f10dfe8b1ea8589685f497766b40b417f9789d Mar 12 00:14:09 crc kubenswrapper[4948]: I0312 00:14:09.938600 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-wb4b8" event={"ID":"567e59f8-2c9e-4332-8eae-ef3e44f5a851","Type":"ContainerStarted","Data":"703cdc8c2a4cd3ea3baeb8515c760c8952b2298b872ac3ff8a80e54d9ddeb457"} Mar 12 00:14:09 crc kubenswrapper[4948]: I0312 00:14:09.938671 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-66df7c8f76-wb4b8" event={"ID":"567e59f8-2c9e-4332-8eae-ef3e44f5a851","Type":"ContainerStarted","Data":"6088259c08ed03828378a8ae62f10dfe8b1ea8589685f497766b40b417f9789d"} Mar 12 00:14:09 crc kubenswrapper[4948]: I0312 00:14:09.938778 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-image-registry/image-registry-66df7c8f76-wb4b8" Mar 12 00:14:09 crc kubenswrapper[4948]: I0312 00:14:09.964960 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-image-registry/image-registry-66df7c8f76-wb4b8" podStartSLOduration=2.964933436 podStartE2EDuration="2.964933436s" podCreationTimestamp="2026-03-12 00:14:07 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:14:09.95920788 +0000 UTC m=+449.414811658" watchObservedRunningTime="2026-03-12 00:14:09.964933436 +0000 UTC m=+449.420537214" Mar 12 00:14:14 crc kubenswrapper[4948]: I0312 00:14:14.785131 4948 patch_prober.go:28] interesting pod/machine-config-daemon-m4xwc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 00:14:14 crc kubenswrapper[4948]: I0312 00:14:14.785716 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" podUID="7e483c2b-08f0-4e92-8e4a-b7281f30af3e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 00:14:16 crc kubenswrapper[4948]: I0312 00:14:16.244995 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-vmmhs" Mar 12 00:14:16 crc kubenswrapper[4948]: I0312 00:14:16.316784 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-vmmhs" Mar 12 00:14:28 crc kubenswrapper[4948]: I0312 00:14:28.296635 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-image-registry/image-registry-66df7c8f76-wb4b8" Mar 12 00:14:28 crc kubenswrapper[4948]: I0312 00:14:28.358548 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-g4kcb"] Mar 12 00:14:44 crc kubenswrapper[4948]: I0312 00:14:44.785636 4948 patch_prober.go:28] interesting pod/machine-config-daemon-m4xwc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 00:14:44 crc kubenswrapper[4948]: I0312 00:14:44.786406 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" podUID="7e483c2b-08f0-4e92-8e4a-b7281f30af3e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 00:14:44 crc kubenswrapper[4948]: I0312 00:14:44.786476 4948 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" Mar 12 00:14:44 crc kubenswrapper[4948]: I0312 00:14:44.787190 4948 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"e9ea8517b452a75f7a88c33d420f90b390e28709d728d5fcba49107a76e86013"} pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 12 00:14:44 crc kubenswrapper[4948]: I0312 00:14:44.787271 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" podUID="7e483c2b-08f0-4e92-8e4a-b7281f30af3e" containerName="machine-config-daemon" containerID="cri-o://e9ea8517b452a75f7a88c33d420f90b390e28709d728d5fcba49107a76e86013" gracePeriod=600 Mar 12 00:14:45 crc kubenswrapper[4948]: I0312 00:14:45.178085 4948 generic.go:334] "Generic (PLEG): container finished" podID="7e483c2b-08f0-4e92-8e4a-b7281f30af3e" containerID="e9ea8517b452a75f7a88c33d420f90b390e28709d728d5fcba49107a76e86013" exitCode=0 Mar 12 00:14:45 crc kubenswrapper[4948]: I0312 00:14:45.178211 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" event={"ID":"7e483c2b-08f0-4e92-8e4a-b7281f30af3e","Type":"ContainerDied","Data":"e9ea8517b452a75f7a88c33d420f90b390e28709d728d5fcba49107a76e86013"} Mar 12 00:14:45 crc kubenswrapper[4948]: I0312 00:14:45.178613 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" event={"ID":"7e483c2b-08f0-4e92-8e4a-b7281f30af3e","Type":"ContainerStarted","Data":"4bae24b53bedc5b7f174f77eb51d6a9cf0630fe1120ad3885ae2181e65d6ec55"} Mar 12 00:14:45 crc kubenswrapper[4948]: I0312 00:14:45.178641 4948 scope.go:117] "RemoveContainer" containerID="82de8f2b633f1b6c295149d077bfd2b308f5acaa429acf411ed16f456536d0ba" Mar 12 00:14:53 crc kubenswrapper[4948]: I0312 00:14:53.415772 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" podUID="af75ce0a-d6bb-48db-a422-08b1d87e1d3f" containerName="registry" containerID="cri-o://3870a7ecb039300e799b9a72a8748670d0c4b779f0a6487f32b8a103b3d173ec" gracePeriod=30 Mar 12 00:14:53 crc kubenswrapper[4948]: I0312 00:14:53.870696 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:14:53 crc kubenswrapper[4948]: I0312 00:14:53.948866 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/af75ce0a-d6bb-48db-a422-08b1d87e1d3f-bound-sa-token\") pod \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " Mar 12 00:14:53 crc kubenswrapper[4948]: I0312 00:14:53.948942 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/af75ce0a-d6bb-48db-a422-08b1d87e1d3f-installation-pull-secrets\") pod \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " Mar 12 00:14:53 crc kubenswrapper[4948]: I0312 00:14:53.948973 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/af75ce0a-d6bb-48db-a422-08b1d87e1d3f-ca-trust-extracted\") pod \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " Mar 12 00:14:53 crc kubenswrapper[4948]: I0312 00:14:53.949000 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/af75ce0a-d6bb-48db-a422-08b1d87e1d3f-registry-tls\") pod \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " Mar 12 00:14:53 crc kubenswrapper[4948]: I0312 00:14:53.949055 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/af75ce0a-d6bb-48db-a422-08b1d87e1d3f-trusted-ca\") pod \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " Mar 12 00:14:53 crc kubenswrapper[4948]: I0312 00:14:53.949135 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jl7d7\" (UniqueName: \"kubernetes.io/projected/af75ce0a-d6bb-48db-a422-08b1d87e1d3f-kube-api-access-jl7d7\") pod \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " Mar 12 00:14:53 crc kubenswrapper[4948]: I0312 00:14:53.949195 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/af75ce0a-d6bb-48db-a422-08b1d87e1d3f-registry-certificates\") pod \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " Mar 12 00:14:53 crc kubenswrapper[4948]: I0312 00:14:53.949371 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"registry-storage\" (UniqueName: \"kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8\") pod \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\" (UID: \"af75ce0a-d6bb-48db-a422-08b1d87e1d3f\") " Mar 12 00:14:53 crc kubenswrapper[4948]: I0312 00:14:53.950617 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af75ce0a-d6bb-48db-a422-08b1d87e1d3f-registry-certificates" (OuterVolumeSpecName: "registry-certificates") pod "af75ce0a-d6bb-48db-a422-08b1d87e1d3f" (UID: "af75ce0a-d6bb-48db-a422-08b1d87e1d3f"). InnerVolumeSpecName "registry-certificates". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:14:53 crc kubenswrapper[4948]: I0312 00:14:53.950497 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/af75ce0a-d6bb-48db-a422-08b1d87e1d3f-trusted-ca" (OuterVolumeSpecName: "trusted-ca") pod "af75ce0a-d6bb-48db-a422-08b1d87e1d3f" (UID: "af75ce0a-d6bb-48db-a422-08b1d87e1d3f"). InnerVolumeSpecName "trusted-ca". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:14:53 crc kubenswrapper[4948]: I0312 00:14:53.958023 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af75ce0a-d6bb-48db-a422-08b1d87e1d3f-kube-api-access-jl7d7" (OuterVolumeSpecName: "kube-api-access-jl7d7") pod "af75ce0a-d6bb-48db-a422-08b1d87e1d3f" (UID: "af75ce0a-d6bb-48db-a422-08b1d87e1d3f"). InnerVolumeSpecName "kube-api-access-jl7d7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:14:53 crc kubenswrapper[4948]: I0312 00:14:53.958497 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/af75ce0a-d6bb-48db-a422-08b1d87e1d3f-installation-pull-secrets" (OuterVolumeSpecName: "installation-pull-secrets") pod "af75ce0a-d6bb-48db-a422-08b1d87e1d3f" (UID: "af75ce0a-d6bb-48db-a422-08b1d87e1d3f"). InnerVolumeSpecName "installation-pull-secrets". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:14:53 crc kubenswrapper[4948]: I0312 00:14:53.958882 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af75ce0a-d6bb-48db-a422-08b1d87e1d3f-bound-sa-token" (OuterVolumeSpecName: "bound-sa-token") pod "af75ce0a-d6bb-48db-a422-08b1d87e1d3f" (UID: "af75ce0a-d6bb-48db-a422-08b1d87e1d3f"). InnerVolumeSpecName "bound-sa-token". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:14:53 crc kubenswrapper[4948]: I0312 00:14:53.960573 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/csi/kubevirt.io.hostpath-provisioner^pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8" (OuterVolumeSpecName: "registry-storage") pod "af75ce0a-d6bb-48db-a422-08b1d87e1d3f" (UID: "af75ce0a-d6bb-48db-a422-08b1d87e1d3f"). InnerVolumeSpecName "pvc-657094db-63f1-4ba8-9a24-edca0e80b7a8". PluginName "kubernetes.io/csi", VolumeGidValue "" Mar 12 00:14:53 crc kubenswrapper[4948]: I0312 00:14:53.962653 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/af75ce0a-d6bb-48db-a422-08b1d87e1d3f-registry-tls" (OuterVolumeSpecName: "registry-tls") pod "af75ce0a-d6bb-48db-a422-08b1d87e1d3f" (UID: "af75ce0a-d6bb-48db-a422-08b1d87e1d3f"). InnerVolumeSpecName "registry-tls". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:14:53 crc kubenswrapper[4948]: I0312 00:14:53.979903 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/af75ce0a-d6bb-48db-a422-08b1d87e1d3f-ca-trust-extracted" (OuterVolumeSpecName: "ca-trust-extracted") pod "af75ce0a-d6bb-48db-a422-08b1d87e1d3f" (UID: "af75ce0a-d6bb-48db-a422-08b1d87e1d3f"). InnerVolumeSpecName "ca-trust-extracted". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 00:14:54 crc kubenswrapper[4948]: I0312 00:14:54.051942 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jl7d7\" (UniqueName: \"kubernetes.io/projected/af75ce0a-d6bb-48db-a422-08b1d87e1d3f-kube-api-access-jl7d7\") on node \"crc\" DevicePath \"\"" Mar 12 00:14:54 crc kubenswrapper[4948]: I0312 00:14:54.052086 4948 reconciler_common.go:293] "Volume detached for volume \"registry-certificates\" (UniqueName: \"kubernetes.io/configmap/af75ce0a-d6bb-48db-a422-08b1d87e1d3f-registry-certificates\") on node \"crc\" DevicePath \"\"" Mar 12 00:14:54 crc kubenswrapper[4948]: I0312 00:14:54.052121 4948 reconciler_common.go:293] "Volume detached for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/af75ce0a-d6bb-48db-a422-08b1d87e1d3f-bound-sa-token\") on node \"crc\" DevicePath \"\"" Mar 12 00:14:54 crc kubenswrapper[4948]: I0312 00:14:54.052151 4948 reconciler_common.go:293] "Volume detached for volume \"registry-tls\" (UniqueName: \"kubernetes.io/projected/af75ce0a-d6bb-48db-a422-08b1d87e1d3f-registry-tls\") on node \"crc\" DevicePath \"\"" Mar 12 00:14:54 crc kubenswrapper[4948]: I0312 00:14:54.052179 4948 reconciler_common.go:293] "Volume detached for volume \"installation-pull-secrets\" (UniqueName: \"kubernetes.io/secret/af75ce0a-d6bb-48db-a422-08b1d87e1d3f-installation-pull-secrets\") on node \"crc\" DevicePath \"\"" Mar 12 00:14:54 crc kubenswrapper[4948]: I0312 00:14:54.052203 4948 reconciler_common.go:293] "Volume detached for volume \"ca-trust-extracted\" (UniqueName: \"kubernetes.io/empty-dir/af75ce0a-d6bb-48db-a422-08b1d87e1d3f-ca-trust-extracted\") on node \"crc\" DevicePath \"\"" Mar 12 00:14:54 crc kubenswrapper[4948]: I0312 00:14:54.052222 4948 reconciler_common.go:293] "Volume detached for volume \"trusted-ca\" (UniqueName: \"kubernetes.io/configmap/af75ce0a-d6bb-48db-a422-08b1d87e1d3f-trusted-ca\") on node \"crc\" DevicePath \"\"" Mar 12 00:14:54 crc kubenswrapper[4948]: I0312 00:14:54.246813 4948 generic.go:334] "Generic (PLEG): container finished" podID="af75ce0a-d6bb-48db-a422-08b1d87e1d3f" containerID="3870a7ecb039300e799b9a72a8748670d0c4b779f0a6487f32b8a103b3d173ec" exitCode=0 Mar 12 00:14:54 crc kubenswrapper[4948]: I0312 00:14:54.246898 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" Mar 12 00:14:54 crc kubenswrapper[4948]: I0312 00:14:54.246883 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" event={"ID":"af75ce0a-d6bb-48db-a422-08b1d87e1d3f","Type":"ContainerDied","Data":"3870a7ecb039300e799b9a72a8748670d0c4b779f0a6487f32b8a103b3d173ec"} Mar 12 00:14:54 crc kubenswrapper[4948]: I0312 00:14:54.247121 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-image-registry/image-registry-697d97f7c8-g4kcb" event={"ID":"af75ce0a-d6bb-48db-a422-08b1d87e1d3f","Type":"ContainerDied","Data":"58ba7b108182e3732e9b992a4f96564f36fb0731b067233c98e570d33cb87f51"} Mar 12 00:14:54 crc kubenswrapper[4948]: I0312 00:14:54.247163 4948 scope.go:117] "RemoveContainer" containerID="3870a7ecb039300e799b9a72a8748670d0c4b779f0a6487f32b8a103b3d173ec" Mar 12 00:14:54 crc kubenswrapper[4948]: I0312 00:14:54.273894 4948 scope.go:117] "RemoveContainer" containerID="3870a7ecb039300e799b9a72a8748670d0c4b779f0a6487f32b8a103b3d173ec" Mar 12 00:14:54 crc kubenswrapper[4948]: E0312 00:14:54.274937 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"3870a7ecb039300e799b9a72a8748670d0c4b779f0a6487f32b8a103b3d173ec\": container with ID starting with 3870a7ecb039300e799b9a72a8748670d0c4b779f0a6487f32b8a103b3d173ec not found: ID does not exist" containerID="3870a7ecb039300e799b9a72a8748670d0c4b779f0a6487f32b8a103b3d173ec" Mar 12 00:14:54 crc kubenswrapper[4948]: I0312 00:14:54.274998 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"3870a7ecb039300e799b9a72a8748670d0c4b779f0a6487f32b8a103b3d173ec"} err="failed to get container status \"3870a7ecb039300e799b9a72a8748670d0c4b779f0a6487f32b8a103b3d173ec\": rpc error: code = NotFound desc = could not find container \"3870a7ecb039300e799b9a72a8748670d0c4b779f0a6487f32b8a103b3d173ec\": container with ID starting with 3870a7ecb039300e799b9a72a8748670d0c4b779f0a6487f32b8a103b3d173ec not found: ID does not exist" Mar 12 00:14:54 crc kubenswrapper[4948]: I0312 00:14:54.303141 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-g4kcb"] Mar 12 00:14:54 crc kubenswrapper[4948]: I0312 00:14:54.314116 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-image-registry/image-registry-697d97f7c8-g4kcb"] Mar 12 00:14:55 crc kubenswrapper[4948]: I0312 00:14:55.326290 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="af75ce0a-d6bb-48db-a422-08b1d87e1d3f" path="/var/lib/kubelet/pods/af75ce0a-d6bb-48db-a422-08b1d87e1d3f/volumes" Mar 12 00:15:00 crc kubenswrapper[4948]: I0312 00:15:00.157776 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29554575-jgbz8"] Mar 12 00:15:00 crc kubenswrapper[4948]: E0312 00:15:00.158775 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="af75ce0a-d6bb-48db-a422-08b1d87e1d3f" containerName="registry" Mar 12 00:15:00 crc kubenswrapper[4948]: I0312 00:15:00.158798 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="af75ce0a-d6bb-48db-a422-08b1d87e1d3f" containerName="registry" Mar 12 00:15:00 crc kubenswrapper[4948]: I0312 00:15:00.158978 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="af75ce0a-d6bb-48db-a422-08b1d87e1d3f" containerName="registry" Mar 12 00:15:00 crc kubenswrapper[4948]: I0312 00:15:00.160729 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29554575-jgbz8" Mar 12 00:15:00 crc kubenswrapper[4948]: I0312 00:15:00.167806 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operator-lifecycle-manager"/"collect-profiles-config" Mar 12 00:15:00 crc kubenswrapper[4948]: I0312 00:15:00.168188 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operator-lifecycle-manager"/"collect-profiles-dockercfg-kzf4t" Mar 12 00:15:00 crc kubenswrapper[4948]: I0312 00:15:00.170440 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29554575-jgbz8"] Mar 12 00:15:00 crc kubenswrapper[4948]: I0312 00:15:00.356096 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5e6cf948-4b38-4386-9eb4-4a67414c5236-secret-volume\") pod \"collect-profiles-29554575-jgbz8\" (UID: \"5e6cf948-4b38-4386-9eb4-4a67414c5236\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29554575-jgbz8" Mar 12 00:15:00 crc kubenswrapper[4948]: I0312 00:15:00.356206 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5e6cf948-4b38-4386-9eb4-4a67414c5236-config-volume\") pod \"collect-profiles-29554575-jgbz8\" (UID: \"5e6cf948-4b38-4386-9eb4-4a67414c5236\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29554575-jgbz8" Mar 12 00:15:00 crc kubenswrapper[4948]: I0312 00:15:00.356648 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-6wcg7\" (UniqueName: \"kubernetes.io/projected/5e6cf948-4b38-4386-9eb4-4a67414c5236-kube-api-access-6wcg7\") pod \"collect-profiles-29554575-jgbz8\" (UID: \"5e6cf948-4b38-4386-9eb4-4a67414c5236\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29554575-jgbz8" Mar 12 00:15:00 crc kubenswrapper[4948]: I0312 00:15:00.458295 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5e6cf948-4b38-4386-9eb4-4a67414c5236-config-volume\") pod \"collect-profiles-29554575-jgbz8\" (UID: \"5e6cf948-4b38-4386-9eb4-4a67414c5236\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29554575-jgbz8" Mar 12 00:15:00 crc kubenswrapper[4948]: I0312 00:15:00.458580 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-6wcg7\" (UniqueName: \"kubernetes.io/projected/5e6cf948-4b38-4386-9eb4-4a67414c5236-kube-api-access-6wcg7\") pod \"collect-profiles-29554575-jgbz8\" (UID: \"5e6cf948-4b38-4386-9eb4-4a67414c5236\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29554575-jgbz8" Mar 12 00:15:00 crc kubenswrapper[4948]: I0312 00:15:00.458692 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5e6cf948-4b38-4386-9eb4-4a67414c5236-secret-volume\") pod \"collect-profiles-29554575-jgbz8\" (UID: \"5e6cf948-4b38-4386-9eb4-4a67414c5236\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29554575-jgbz8" Mar 12 00:15:00 crc kubenswrapper[4948]: I0312 00:15:00.462623 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5e6cf948-4b38-4386-9eb4-4a67414c5236-config-volume\") pod \"collect-profiles-29554575-jgbz8\" (UID: \"5e6cf948-4b38-4386-9eb4-4a67414c5236\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29554575-jgbz8" Mar 12 00:15:00 crc kubenswrapper[4948]: I0312 00:15:00.470902 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5e6cf948-4b38-4386-9eb4-4a67414c5236-secret-volume\") pod \"collect-profiles-29554575-jgbz8\" (UID: \"5e6cf948-4b38-4386-9eb4-4a67414c5236\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29554575-jgbz8" Mar 12 00:15:00 crc kubenswrapper[4948]: I0312 00:15:00.493021 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-6wcg7\" (UniqueName: \"kubernetes.io/projected/5e6cf948-4b38-4386-9eb4-4a67414c5236-kube-api-access-6wcg7\") pod \"collect-profiles-29554575-jgbz8\" (UID: \"5e6cf948-4b38-4386-9eb4-4a67414c5236\") " pod="openshift-operator-lifecycle-manager/collect-profiles-29554575-jgbz8" Mar 12 00:15:00 crc kubenswrapper[4948]: I0312 00:15:00.791598 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29554575-jgbz8" Mar 12 00:15:01 crc kubenswrapper[4948]: I0312 00:15:01.029711 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operator-lifecycle-manager/collect-profiles-29554575-jgbz8"] Mar 12 00:15:01 crc kubenswrapper[4948]: I0312 00:15:01.300403 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29554575-jgbz8" event={"ID":"5e6cf948-4b38-4386-9eb4-4a67414c5236","Type":"ContainerStarted","Data":"c41a0aeb011c7536837c6c84c78f5b048039dc27f4e9577d579031fd0bf71fe5"} Mar 12 00:15:01 crc kubenswrapper[4948]: I0312 00:15:01.300765 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29554575-jgbz8" event={"ID":"5e6cf948-4b38-4386-9eb4-4a67414c5236","Type":"ContainerStarted","Data":"c02a4e15caf7090c92b53b575703e0e059c9376971b838079188bd766af18f45"} Mar 12 00:15:01 crc kubenswrapper[4948]: I0312 00:15:01.333787 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operator-lifecycle-manager/collect-profiles-29554575-jgbz8" podStartSLOduration=1.333759467 podStartE2EDuration="1.333759467s" podCreationTimestamp="2026-03-12 00:15:00 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:15:01.328503926 +0000 UTC m=+500.784107694" watchObservedRunningTime="2026-03-12 00:15:01.333759467 +0000 UTC m=+500.789363245" Mar 12 00:15:02 crc kubenswrapper[4948]: I0312 00:15:02.310371 4948 generic.go:334] "Generic (PLEG): container finished" podID="5e6cf948-4b38-4386-9eb4-4a67414c5236" containerID="c41a0aeb011c7536837c6c84c78f5b048039dc27f4e9577d579031fd0bf71fe5" exitCode=0 Mar 12 00:15:02 crc kubenswrapper[4948]: I0312 00:15:02.310429 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29554575-jgbz8" event={"ID":"5e6cf948-4b38-4386-9eb4-4a67414c5236","Type":"ContainerDied","Data":"c41a0aeb011c7536837c6c84c78f5b048039dc27f4e9577d579031fd0bf71fe5"} Mar 12 00:15:03 crc kubenswrapper[4948]: I0312 00:15:03.615501 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29554575-jgbz8" Mar 12 00:15:03 crc kubenswrapper[4948]: I0312 00:15:03.799394 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5e6cf948-4b38-4386-9eb4-4a67414c5236-config-volume\") pod \"5e6cf948-4b38-4386-9eb4-4a67414c5236\" (UID: \"5e6cf948-4b38-4386-9eb4-4a67414c5236\") " Mar 12 00:15:03 crc kubenswrapper[4948]: I0312 00:15:03.799754 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-6wcg7\" (UniqueName: \"kubernetes.io/projected/5e6cf948-4b38-4386-9eb4-4a67414c5236-kube-api-access-6wcg7\") pod \"5e6cf948-4b38-4386-9eb4-4a67414c5236\" (UID: \"5e6cf948-4b38-4386-9eb4-4a67414c5236\") " Mar 12 00:15:03 crc kubenswrapper[4948]: I0312 00:15:03.799856 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5e6cf948-4b38-4386-9eb4-4a67414c5236-secret-volume\") pod \"5e6cf948-4b38-4386-9eb4-4a67414c5236\" (UID: \"5e6cf948-4b38-4386-9eb4-4a67414c5236\") " Mar 12 00:15:03 crc kubenswrapper[4948]: I0312 00:15:03.800409 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/5e6cf948-4b38-4386-9eb4-4a67414c5236-config-volume" (OuterVolumeSpecName: "config-volume") pod "5e6cf948-4b38-4386-9eb4-4a67414c5236" (UID: "5e6cf948-4b38-4386-9eb4-4a67414c5236"). InnerVolumeSpecName "config-volume". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:15:03 crc kubenswrapper[4948]: I0312 00:15:03.810224 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/5e6cf948-4b38-4386-9eb4-4a67414c5236-secret-volume" (OuterVolumeSpecName: "secret-volume") pod "5e6cf948-4b38-4386-9eb4-4a67414c5236" (UID: "5e6cf948-4b38-4386-9eb4-4a67414c5236"). InnerVolumeSpecName "secret-volume". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:15:03 crc kubenswrapper[4948]: I0312 00:15:03.811330 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/5e6cf948-4b38-4386-9eb4-4a67414c5236-kube-api-access-6wcg7" (OuterVolumeSpecName: "kube-api-access-6wcg7") pod "5e6cf948-4b38-4386-9eb4-4a67414c5236" (UID: "5e6cf948-4b38-4386-9eb4-4a67414c5236"). InnerVolumeSpecName "kube-api-access-6wcg7". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:15:03 crc kubenswrapper[4948]: I0312 00:15:03.901470 4948 reconciler_common.go:293] "Volume detached for volume \"config-volume\" (UniqueName: \"kubernetes.io/configmap/5e6cf948-4b38-4386-9eb4-4a67414c5236-config-volume\") on node \"crc\" DevicePath \"\"" Mar 12 00:15:03 crc kubenswrapper[4948]: I0312 00:15:03.901506 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-6wcg7\" (UniqueName: \"kubernetes.io/projected/5e6cf948-4b38-4386-9eb4-4a67414c5236-kube-api-access-6wcg7\") on node \"crc\" DevicePath \"\"" Mar 12 00:15:03 crc kubenswrapper[4948]: I0312 00:15:03.901521 4948 reconciler_common.go:293] "Volume detached for volume \"secret-volume\" (UniqueName: \"kubernetes.io/secret/5e6cf948-4b38-4386-9eb4-4a67414c5236-secret-volume\") on node \"crc\" DevicePath \"\"" Mar 12 00:15:04 crc kubenswrapper[4948]: I0312 00:15:04.324140 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operator-lifecycle-manager/collect-profiles-29554575-jgbz8" event={"ID":"5e6cf948-4b38-4386-9eb4-4a67414c5236","Type":"ContainerDied","Data":"c02a4e15caf7090c92b53b575703e0e059c9376971b838079188bd766af18f45"} Mar 12 00:15:04 crc kubenswrapper[4948]: I0312 00:15:04.324196 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c02a4e15caf7090c92b53b575703e0e059c9376971b838079188bd766af18f45" Mar 12 00:15:04 crc kubenswrapper[4948]: I0312 00:15:04.324206 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-operator-lifecycle-manager/collect-profiles-29554575-jgbz8" Mar 12 00:16:00 crc kubenswrapper[4948]: I0312 00:16:00.150048 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29554576-k9gx6"] Mar 12 00:16:00 crc kubenswrapper[4948]: E0312 00:16:00.157712 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="5e6cf948-4b38-4386-9eb4-4a67414c5236" containerName="collect-profiles" Mar 12 00:16:00 crc kubenswrapper[4948]: I0312 00:16:00.157756 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="5e6cf948-4b38-4386-9eb4-4a67414c5236" containerName="collect-profiles" Mar 12 00:16:00 crc kubenswrapper[4948]: I0312 00:16:00.158057 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="5e6cf948-4b38-4386-9eb4-4a67414c5236" containerName="collect-profiles" Mar 12 00:16:00 crc kubenswrapper[4948]: I0312 00:16:00.159915 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29554576-k9gx6" Mar 12 00:16:00 crc kubenswrapper[4948]: I0312 00:16:00.164049 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 00:16:00 crc kubenswrapper[4948]: I0312 00:16:00.164212 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 00:16:00 crc kubenswrapper[4948]: I0312 00:16:00.164065 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-44g5m" Mar 12 00:16:00 crc kubenswrapper[4948]: I0312 00:16:00.172635 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29554576-k9gx6"] Mar 12 00:16:00 crc kubenswrapper[4948]: I0312 00:16:00.329161 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-rq9cq\" (UniqueName: \"kubernetes.io/projected/551726c9-76b6-435a-ae73-c43821a95491-kube-api-access-rq9cq\") pod \"auto-csr-approver-29554576-k9gx6\" (UID: \"551726c9-76b6-435a-ae73-c43821a95491\") " pod="openshift-infra/auto-csr-approver-29554576-k9gx6" Mar 12 00:16:00 crc kubenswrapper[4948]: I0312 00:16:00.430502 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-rq9cq\" (UniqueName: \"kubernetes.io/projected/551726c9-76b6-435a-ae73-c43821a95491-kube-api-access-rq9cq\") pod \"auto-csr-approver-29554576-k9gx6\" (UID: \"551726c9-76b6-435a-ae73-c43821a95491\") " pod="openshift-infra/auto-csr-approver-29554576-k9gx6" Mar 12 00:16:00 crc kubenswrapper[4948]: I0312 00:16:00.457940 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-rq9cq\" (UniqueName: \"kubernetes.io/projected/551726c9-76b6-435a-ae73-c43821a95491-kube-api-access-rq9cq\") pod \"auto-csr-approver-29554576-k9gx6\" (UID: \"551726c9-76b6-435a-ae73-c43821a95491\") " pod="openshift-infra/auto-csr-approver-29554576-k9gx6" Mar 12 00:16:00 crc kubenswrapper[4948]: I0312 00:16:00.489357 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29554576-k9gx6" Mar 12 00:16:00 crc kubenswrapper[4948]: I0312 00:16:00.764935 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29554576-k9gx6"] Mar 12 00:16:00 crc kubenswrapper[4948]: I0312 00:16:00.776216 4948 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 12 00:16:01 crc kubenswrapper[4948]: I0312 00:16:01.762732 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29554576-k9gx6" event={"ID":"551726c9-76b6-435a-ae73-c43821a95491","Type":"ContainerStarted","Data":"fa2fbf53496c6d73bc4c96ecf5c3f686974c11ddde18c43da6dcadfe535cdd7d"} Mar 12 00:16:02 crc kubenswrapper[4948]: I0312 00:16:02.776383 4948 generic.go:334] "Generic (PLEG): container finished" podID="551726c9-76b6-435a-ae73-c43821a95491" containerID="cd53f999f5fd5a2b763323f1d841e6ceed15b820af87bb68737992dda04f42c2" exitCode=0 Mar 12 00:16:02 crc kubenswrapper[4948]: I0312 00:16:02.776453 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29554576-k9gx6" event={"ID":"551726c9-76b6-435a-ae73-c43821a95491","Type":"ContainerDied","Data":"cd53f999f5fd5a2b763323f1d841e6ceed15b820af87bb68737992dda04f42c2"} Mar 12 00:16:04 crc kubenswrapper[4948]: I0312 00:16:04.056485 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29554576-k9gx6" Mar 12 00:16:04 crc kubenswrapper[4948]: I0312 00:16:04.073979 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-rq9cq\" (UniqueName: \"kubernetes.io/projected/551726c9-76b6-435a-ae73-c43821a95491-kube-api-access-rq9cq\") pod \"551726c9-76b6-435a-ae73-c43821a95491\" (UID: \"551726c9-76b6-435a-ae73-c43821a95491\") " Mar 12 00:16:04 crc kubenswrapper[4948]: I0312 00:16:04.119612 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/551726c9-76b6-435a-ae73-c43821a95491-kube-api-access-rq9cq" (OuterVolumeSpecName: "kube-api-access-rq9cq") pod "551726c9-76b6-435a-ae73-c43821a95491" (UID: "551726c9-76b6-435a-ae73-c43821a95491"). InnerVolumeSpecName "kube-api-access-rq9cq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:16:04 crc kubenswrapper[4948]: I0312 00:16:04.175245 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-rq9cq\" (UniqueName: \"kubernetes.io/projected/551726c9-76b6-435a-ae73-c43821a95491-kube-api-access-rq9cq\") on node \"crc\" DevicePath \"\"" Mar 12 00:16:04 crc kubenswrapper[4948]: I0312 00:16:04.790463 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29554576-k9gx6" event={"ID":"551726c9-76b6-435a-ae73-c43821a95491","Type":"ContainerDied","Data":"fa2fbf53496c6d73bc4c96ecf5c3f686974c11ddde18c43da6dcadfe535cdd7d"} Mar 12 00:16:04 crc kubenswrapper[4948]: I0312 00:16:04.790739 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fa2fbf53496c6d73bc4c96ecf5c3f686974c11ddde18c43da6dcadfe535cdd7d" Mar 12 00:16:04 crc kubenswrapper[4948]: I0312 00:16:04.790549 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29554576-k9gx6" Mar 12 00:16:05 crc kubenswrapper[4948]: I0312 00:16:05.135582 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29554570-2b2hb"] Mar 12 00:16:05 crc kubenswrapper[4948]: I0312 00:16:05.142170 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29554570-2b2hb"] Mar 12 00:16:05 crc kubenswrapper[4948]: I0312 00:16:05.325641 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2fb324d6-44d4-440d-ae60-a5356a3a057f" path="/var/lib/kubelet/pods/2fb324d6-44d4-440d-ae60-a5356a3a057f/volumes" Mar 12 00:16:56 crc kubenswrapper[4948]: I0312 00:16:56.685434 4948 scope.go:117] "RemoveContainer" containerID="9cece2d83a0870b294348ea430867de0e80ef150fd12302e1925a21396eb90d1" Mar 12 00:17:14 crc kubenswrapper[4948]: I0312 00:17:14.784809 4948 patch_prober.go:28] interesting pod/machine-config-daemon-m4xwc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 00:17:14 crc kubenswrapper[4948]: I0312 00:17:14.785605 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" podUID="7e483c2b-08f0-4e92-8e4a-b7281f30af3e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 00:17:44 crc kubenswrapper[4948]: I0312 00:17:44.785348 4948 patch_prober.go:28] interesting pod/machine-config-daemon-m4xwc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 00:17:44 crc kubenswrapper[4948]: I0312 00:17:44.786105 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" podUID="7e483c2b-08f0-4e92-8e4a-b7281f30af3e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 00:17:56 crc kubenswrapper[4948]: I0312 00:17:56.724755 4948 scope.go:117] "RemoveContainer" containerID="05fd166aeeb70277dab1df229c8cbde7b85139a276d0732903b8b28713344011" Mar 12 00:17:56 crc kubenswrapper[4948]: I0312 00:17:56.750578 4948 scope.go:117] "RemoveContainer" containerID="e27ca11906dfd90f4412b456fd32a7dce40145941516de818777da1c7d83353c" Mar 12 00:17:56 crc kubenswrapper[4948]: I0312 00:17:56.771261 4948 scope.go:117] "RemoveContainer" containerID="6c842526274736c2e138ac271848bc99ec3ab0bb8912406eba8443b61f20cc89" Mar 12 00:18:00 crc kubenswrapper[4948]: I0312 00:18:00.148793 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29554578-fz9lc"] Mar 12 00:18:00 crc kubenswrapper[4948]: E0312 00:18:00.149512 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="551726c9-76b6-435a-ae73-c43821a95491" containerName="oc" Mar 12 00:18:00 crc kubenswrapper[4948]: I0312 00:18:00.149537 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="551726c9-76b6-435a-ae73-c43821a95491" containerName="oc" Mar 12 00:18:00 crc kubenswrapper[4948]: I0312 00:18:00.149700 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="551726c9-76b6-435a-ae73-c43821a95491" containerName="oc" Mar 12 00:18:00 crc kubenswrapper[4948]: I0312 00:18:00.150470 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29554578-fz9lc" Mar 12 00:18:00 crc kubenswrapper[4948]: I0312 00:18:00.159979 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-44g5m" Mar 12 00:18:00 crc kubenswrapper[4948]: I0312 00:18:00.160460 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 00:18:00 crc kubenswrapper[4948]: I0312 00:18:00.160507 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 00:18:00 crc kubenswrapper[4948]: I0312 00:18:00.162399 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29554578-fz9lc"] Mar 12 00:18:00 crc kubenswrapper[4948]: I0312 00:18:00.276284 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-x5r9p\" (UniqueName: \"kubernetes.io/projected/3b3f134e-8447-45bb-9b8e-ab8980e7a614-kube-api-access-x5r9p\") pod \"auto-csr-approver-29554578-fz9lc\" (UID: \"3b3f134e-8447-45bb-9b8e-ab8980e7a614\") " pod="openshift-infra/auto-csr-approver-29554578-fz9lc" Mar 12 00:18:00 crc kubenswrapper[4948]: I0312 00:18:00.378155 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-x5r9p\" (UniqueName: \"kubernetes.io/projected/3b3f134e-8447-45bb-9b8e-ab8980e7a614-kube-api-access-x5r9p\") pod \"auto-csr-approver-29554578-fz9lc\" (UID: \"3b3f134e-8447-45bb-9b8e-ab8980e7a614\") " pod="openshift-infra/auto-csr-approver-29554578-fz9lc" Mar 12 00:18:00 crc kubenswrapper[4948]: I0312 00:18:00.424679 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-x5r9p\" (UniqueName: \"kubernetes.io/projected/3b3f134e-8447-45bb-9b8e-ab8980e7a614-kube-api-access-x5r9p\") pod \"auto-csr-approver-29554578-fz9lc\" (UID: \"3b3f134e-8447-45bb-9b8e-ab8980e7a614\") " pod="openshift-infra/auto-csr-approver-29554578-fz9lc" Mar 12 00:18:00 crc kubenswrapper[4948]: I0312 00:18:00.493724 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29554578-fz9lc" Mar 12 00:18:00 crc kubenswrapper[4948]: I0312 00:18:00.802868 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29554578-fz9lc"] Mar 12 00:18:01 crc kubenswrapper[4948]: I0312 00:18:01.669801 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29554578-fz9lc" event={"ID":"3b3f134e-8447-45bb-9b8e-ab8980e7a614","Type":"ContainerStarted","Data":"cedef6c8a999ac6ad95fe58f8e3ce52f5d6ceda573682120f357a18fe2b7418d"} Mar 12 00:18:02 crc kubenswrapper[4948]: I0312 00:18:02.680705 4948 generic.go:334] "Generic (PLEG): container finished" podID="3b3f134e-8447-45bb-9b8e-ab8980e7a614" containerID="c4d3d32ce81f8386f0b5bf4319cfb711a12429b17af996353ce7b07372b881af" exitCode=0 Mar 12 00:18:02 crc kubenswrapper[4948]: I0312 00:18:02.680780 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29554578-fz9lc" event={"ID":"3b3f134e-8447-45bb-9b8e-ab8980e7a614","Type":"ContainerDied","Data":"c4d3d32ce81f8386f0b5bf4319cfb711a12429b17af996353ce7b07372b881af"} Mar 12 00:18:03 crc kubenswrapper[4948]: I0312 00:18:03.963998 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29554578-fz9lc" Mar 12 00:18:04 crc kubenswrapper[4948]: I0312 00:18:04.032026 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-x5r9p\" (UniqueName: \"kubernetes.io/projected/3b3f134e-8447-45bb-9b8e-ab8980e7a614-kube-api-access-x5r9p\") pod \"3b3f134e-8447-45bb-9b8e-ab8980e7a614\" (UID: \"3b3f134e-8447-45bb-9b8e-ab8980e7a614\") " Mar 12 00:18:04 crc kubenswrapper[4948]: I0312 00:18:04.042564 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3b3f134e-8447-45bb-9b8e-ab8980e7a614-kube-api-access-x5r9p" (OuterVolumeSpecName: "kube-api-access-x5r9p") pod "3b3f134e-8447-45bb-9b8e-ab8980e7a614" (UID: "3b3f134e-8447-45bb-9b8e-ab8980e7a614"). InnerVolumeSpecName "kube-api-access-x5r9p". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:18:04 crc kubenswrapper[4948]: I0312 00:18:04.133858 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-x5r9p\" (UniqueName: \"kubernetes.io/projected/3b3f134e-8447-45bb-9b8e-ab8980e7a614-kube-api-access-x5r9p\") on node \"crc\" DevicePath \"\"" Mar 12 00:18:04 crc kubenswrapper[4948]: I0312 00:18:04.705098 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29554578-fz9lc" Mar 12 00:18:04 crc kubenswrapper[4948]: I0312 00:18:04.705581 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29554578-fz9lc" event={"ID":"3b3f134e-8447-45bb-9b8e-ab8980e7a614","Type":"ContainerDied","Data":"cedef6c8a999ac6ad95fe58f8e3ce52f5d6ceda573682120f357a18fe2b7418d"} Mar 12 00:18:04 crc kubenswrapper[4948]: I0312 00:18:04.705640 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="cedef6c8a999ac6ad95fe58f8e3ce52f5d6ceda573682120f357a18fe2b7418d" Mar 12 00:18:05 crc kubenswrapper[4948]: I0312 00:18:05.025494 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29554572-szktg"] Mar 12 00:18:05 crc kubenswrapper[4948]: I0312 00:18:05.028422 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29554572-szktg"] Mar 12 00:18:05 crc kubenswrapper[4948]: I0312 00:18:05.327392 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="7401d71e-946e-4b13-b187-13db4b1be35b" path="/var/lib/kubelet/pods/7401d71e-946e-4b13-b187-13db4b1be35b/volumes" Mar 12 00:18:14 crc kubenswrapper[4948]: I0312 00:18:14.784709 4948 patch_prober.go:28] interesting pod/machine-config-daemon-m4xwc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 00:18:14 crc kubenswrapper[4948]: I0312 00:18:14.785668 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" podUID="7e483c2b-08f0-4e92-8e4a-b7281f30af3e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 00:18:14 crc kubenswrapper[4948]: I0312 00:18:14.785770 4948 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" Mar 12 00:18:14 crc kubenswrapper[4948]: I0312 00:18:14.787202 4948 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4bae24b53bedc5b7f174f77eb51d6a9cf0630fe1120ad3885ae2181e65d6ec55"} pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 12 00:18:14 crc kubenswrapper[4948]: I0312 00:18:14.787363 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" podUID="7e483c2b-08f0-4e92-8e4a-b7281f30af3e" containerName="machine-config-daemon" containerID="cri-o://4bae24b53bedc5b7f174f77eb51d6a9cf0630fe1120ad3885ae2181e65d6ec55" gracePeriod=600 Mar 12 00:18:15 crc kubenswrapper[4948]: I0312 00:18:15.781098 4948 generic.go:334] "Generic (PLEG): container finished" podID="7e483c2b-08f0-4e92-8e4a-b7281f30af3e" containerID="4bae24b53bedc5b7f174f77eb51d6a9cf0630fe1120ad3885ae2181e65d6ec55" exitCode=0 Mar 12 00:18:15 crc kubenswrapper[4948]: I0312 00:18:15.781181 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" event={"ID":"7e483c2b-08f0-4e92-8e4a-b7281f30af3e","Type":"ContainerDied","Data":"4bae24b53bedc5b7f174f77eb51d6a9cf0630fe1120ad3885ae2181e65d6ec55"} Mar 12 00:18:15 crc kubenswrapper[4948]: I0312 00:18:15.781719 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" event={"ID":"7e483c2b-08f0-4e92-8e4a-b7281f30af3e","Type":"ContainerStarted","Data":"d57d486c11ea6dd0bb65ccd254f325841bcd048227a6896a9722dd323a6190db"} Mar 12 00:18:15 crc kubenswrapper[4948]: I0312 00:18:15.781751 4948 scope.go:117] "RemoveContainer" containerID="e9ea8517b452a75f7a88c33d420f90b390e28709d728d5fcba49107a76e86013" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.144621 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-f7mp5"] Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.147206 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" podUID="15117613-9fad-48c7-98c4-a2d84502ded9" containerName="nbdb" containerID="cri-o://2e481ace9e84b6aa66cbdda3ee8425c4120a665915d2c70ae4faf60bba6205e2" gracePeriod=30 Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.147521 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" podUID="15117613-9fad-48c7-98c4-a2d84502ded9" containerName="kube-rbac-proxy-ovn-metrics" containerID="cri-o://84af4cdd213d90636a889a8a8a3aaca3dee1e2677368efa0ffc768940c2cc994" gracePeriod=30 Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.147705 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" podUID="15117613-9fad-48c7-98c4-a2d84502ded9" containerName="ovn-acl-logging" containerID="cri-o://ec9d5e77753fda70350353daa0967257dd4929dba4440a95495ab4c2a0659aef" gracePeriod=30 Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.147396 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" podUID="15117613-9fad-48c7-98c4-a2d84502ded9" containerName="sbdb" containerID="cri-o://46fa6140f4f4bd32e298fa36f6d7af6890029f9bc3080d2b4d94be8307346de8" gracePeriod=30 Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.147617 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" podUID="15117613-9fad-48c7-98c4-a2d84502ded9" containerName="kube-rbac-proxy-node" containerID="cri-o://d4036d7131db00afd8eab8e093626b688297dc776b5af80c3c38b51da7124085" gracePeriod=30 Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.147572 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" podUID="15117613-9fad-48c7-98c4-a2d84502ded9" containerName="northd" containerID="cri-o://c9eadefd0babc119f776d48c9d2c9e4bced58db685af9a8e97f25e6a9a2e3c89" gracePeriod=30 Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.154280 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" podUID="15117613-9fad-48c7-98c4-a2d84502ded9" containerName="ovn-controller" containerID="cri-o://8cad943edb9a8d445b33a6e70d85b475c1ccba460a6056bccae94fe60fd6b73f" gracePeriod=30 Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.214927 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" podUID="15117613-9fad-48c7-98c4-a2d84502ded9" containerName="ovnkube-controller" containerID="cri-o://d3de0f7f4aa39ee28267d3a0041033e89da7e51713f417910e09007110b7b9c2" gracePeriod=30 Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.507017 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f7mp5_15117613-9fad-48c7-98c4-a2d84502ded9/ovnkube-controller/3.log" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.510661 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f7mp5_15117613-9fad-48c7-98c4-a2d84502ded9/ovn-acl-logging/0.log" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.511460 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f7mp5_15117613-9fad-48c7-98c4-a2d84502ded9/ovn-controller/0.log" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.512015 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.598617 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-lqrtr"] Mar 12 00:18:38 crc kubenswrapper[4948]: E0312 00:18:38.598923 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15117613-9fad-48c7-98c4-a2d84502ded9" containerName="nbdb" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.598952 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="15117613-9fad-48c7-98c4-a2d84502ded9" containerName="nbdb" Mar 12 00:18:38 crc kubenswrapper[4948]: E0312 00:18:38.598980 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15117613-9fad-48c7-98c4-a2d84502ded9" containerName="sbdb" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.598994 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="15117613-9fad-48c7-98c4-a2d84502ded9" containerName="sbdb" Mar 12 00:18:38 crc kubenswrapper[4948]: E0312 00:18:38.599013 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15117613-9fad-48c7-98c4-a2d84502ded9" containerName="ovnkube-controller" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.599027 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="15117613-9fad-48c7-98c4-a2d84502ded9" containerName="ovnkube-controller" Mar 12 00:18:38 crc kubenswrapper[4948]: E0312 00:18:38.599045 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15117613-9fad-48c7-98c4-a2d84502ded9" containerName="ovnkube-controller" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.599058 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="15117613-9fad-48c7-98c4-a2d84502ded9" containerName="ovnkube-controller" Mar 12 00:18:38 crc kubenswrapper[4948]: E0312 00:18:38.599073 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3b3f134e-8447-45bb-9b8e-ab8980e7a614" containerName="oc" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.599086 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="3b3f134e-8447-45bb-9b8e-ab8980e7a614" containerName="oc" Mar 12 00:18:38 crc kubenswrapper[4948]: E0312 00:18:38.599100 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15117613-9fad-48c7-98c4-a2d84502ded9" containerName="northd" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.599113 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="15117613-9fad-48c7-98c4-a2d84502ded9" containerName="northd" Mar 12 00:18:38 crc kubenswrapper[4948]: E0312 00:18:38.599135 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15117613-9fad-48c7-98c4-a2d84502ded9" containerName="ovnkube-controller" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.599148 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="15117613-9fad-48c7-98c4-a2d84502ded9" containerName="ovnkube-controller" Mar 12 00:18:38 crc kubenswrapper[4948]: E0312 00:18:38.599164 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15117613-9fad-48c7-98c4-a2d84502ded9" containerName="ovn-controller" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.599177 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="15117613-9fad-48c7-98c4-a2d84502ded9" containerName="ovn-controller" Mar 12 00:18:38 crc kubenswrapper[4948]: E0312 00:18:38.599196 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15117613-9fad-48c7-98c4-a2d84502ded9" containerName="ovn-acl-logging" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.599210 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="15117613-9fad-48c7-98c4-a2d84502ded9" containerName="ovn-acl-logging" Mar 12 00:18:38 crc kubenswrapper[4948]: E0312 00:18:38.599231 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15117613-9fad-48c7-98c4-a2d84502ded9" containerName="kubecfg-setup" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.599244 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="15117613-9fad-48c7-98c4-a2d84502ded9" containerName="kubecfg-setup" Mar 12 00:18:38 crc kubenswrapper[4948]: E0312 00:18:38.599405 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15117613-9fad-48c7-98c4-a2d84502ded9" containerName="kube-rbac-proxy-node" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.599423 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="15117613-9fad-48c7-98c4-a2d84502ded9" containerName="kube-rbac-proxy-node" Mar 12 00:18:38 crc kubenswrapper[4948]: E0312 00:18:38.599441 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15117613-9fad-48c7-98c4-a2d84502ded9" containerName="kube-rbac-proxy-ovn-metrics" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.599453 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="15117613-9fad-48c7-98c4-a2d84502ded9" containerName="kube-rbac-proxy-ovn-metrics" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.599622 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="15117613-9fad-48c7-98c4-a2d84502ded9" containerName="sbdb" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.599638 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="15117613-9fad-48c7-98c4-a2d84502ded9" containerName="ovnkube-controller" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.599657 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="15117613-9fad-48c7-98c4-a2d84502ded9" containerName="ovn-acl-logging" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.599675 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="15117613-9fad-48c7-98c4-a2d84502ded9" containerName="ovn-controller" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.599692 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="15117613-9fad-48c7-98c4-a2d84502ded9" containerName="nbdb" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.599715 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="3b3f134e-8447-45bb-9b8e-ab8980e7a614" containerName="oc" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.599733 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="15117613-9fad-48c7-98c4-a2d84502ded9" containerName="kube-rbac-proxy-ovn-metrics" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.599749 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="15117613-9fad-48c7-98c4-a2d84502ded9" containerName="ovnkube-controller" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.599762 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="15117613-9fad-48c7-98c4-a2d84502ded9" containerName="ovnkube-controller" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.599780 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="15117613-9fad-48c7-98c4-a2d84502ded9" containerName="northd" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.599797 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="15117613-9fad-48c7-98c4-a2d84502ded9" containerName="kube-rbac-proxy-node" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.599813 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="15117613-9fad-48c7-98c4-a2d84502ded9" containerName="ovnkube-controller" Mar 12 00:18:38 crc kubenswrapper[4948]: E0312 00:18:38.599965 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15117613-9fad-48c7-98c4-a2d84502ded9" containerName="ovnkube-controller" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.599980 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="15117613-9fad-48c7-98c4-a2d84502ded9" containerName="ovnkube-controller" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.600163 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="15117613-9fad-48c7-98c4-a2d84502ded9" containerName="ovnkube-controller" Mar 12 00:18:38 crc kubenswrapper[4948]: E0312 00:18:38.600427 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="15117613-9fad-48c7-98c4-a2d84502ded9" containerName="ovnkube-controller" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.600454 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="15117613-9fad-48c7-98c4-a2d84502ded9" containerName="ovnkube-controller" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.603394 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.630191 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-host-cni-netd\") pod \"15117613-9fad-48c7-98c4-a2d84502ded9\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.630251 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-host-cni-bin\") pod \"15117613-9fad-48c7-98c4-a2d84502ded9\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.630288 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-log-socket\") pod \"15117613-9fad-48c7-98c4-a2d84502ded9\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.630328 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-host-cni-netd" (OuterVolumeSpecName: "host-cni-netd") pod "15117613-9fad-48c7-98c4-a2d84502ded9" (UID: "15117613-9fad-48c7-98c4-a2d84502ded9"). InnerVolumeSpecName "host-cni-netd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.630358 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dvwhv\" (UniqueName: \"kubernetes.io/projected/15117613-9fad-48c7-98c4-a2d84502ded9-kube-api-access-dvwhv\") pod \"15117613-9fad-48c7-98c4-a2d84502ded9\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.630383 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-host-cni-bin" (OuterVolumeSpecName: "host-cni-bin") pod "15117613-9fad-48c7-98c4-a2d84502ded9" (UID: "15117613-9fad-48c7-98c4-a2d84502ded9"). InnerVolumeSpecName "host-cni-bin". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.630410 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-log-socket" (OuterVolumeSpecName: "log-socket") pod "15117613-9fad-48c7-98c4-a2d84502ded9" (UID: "15117613-9fad-48c7-98c4-a2d84502ded9"). InnerVolumeSpecName "log-socket". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.630506 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/15117613-9fad-48c7-98c4-a2d84502ded9-ovnkube-config\") pod \"15117613-9fad-48c7-98c4-a2d84502ded9\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.630624 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-host-var-lib-cni-networks-ovn-kubernetes\") pod \"15117613-9fad-48c7-98c4-a2d84502ded9\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.630659 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-host-var-lib-cni-networks-ovn-kubernetes" (OuterVolumeSpecName: "host-var-lib-cni-networks-ovn-kubernetes") pod "15117613-9fad-48c7-98c4-a2d84502ded9" (UID: "15117613-9fad-48c7-98c4-a2d84502ded9"). InnerVolumeSpecName "host-var-lib-cni-networks-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.630885 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-run-openvswitch\") pod \"15117613-9fad-48c7-98c4-a2d84502ded9\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.630916 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-run-openvswitch" (OuterVolumeSpecName: "run-openvswitch") pod "15117613-9fad-48c7-98c4-a2d84502ded9" (UID: "15117613-9fad-48c7-98c4-a2d84502ded9"). InnerVolumeSpecName "run-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.631120 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-etc-openvswitch\") pod \"15117613-9fad-48c7-98c4-a2d84502ded9\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.631182 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-etc-openvswitch" (OuterVolumeSpecName: "etc-openvswitch") pod "15117613-9fad-48c7-98c4-a2d84502ded9" (UID: "15117613-9fad-48c7-98c4-a2d84502ded9"). InnerVolumeSpecName "etc-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.631260 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-node-log\") pod \"15117613-9fad-48c7-98c4-a2d84502ded9\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.631398 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-run-systemd\") pod \"15117613-9fad-48c7-98c4-a2d84502ded9\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.631477 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-run-ovn\") pod \"15117613-9fad-48c7-98c4-a2d84502ded9\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.631565 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-var-lib-openvswitch\") pod \"15117613-9fad-48c7-98c4-a2d84502ded9\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.631644 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-host-run-netns\") pod \"15117613-9fad-48c7-98c4-a2d84502ded9\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.631342 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-node-log" (OuterVolumeSpecName: "node-log") pod "15117613-9fad-48c7-98c4-a2d84502ded9" (UID: "15117613-9fad-48c7-98c4-a2d84502ded9"). InnerVolumeSpecName "node-log". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.631583 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-run-ovn" (OuterVolumeSpecName: "run-ovn") pod "15117613-9fad-48c7-98c4-a2d84502ded9" (UID: "15117613-9fad-48c7-98c4-a2d84502ded9"). InnerVolumeSpecName "run-ovn". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.631690 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-host-run-netns" (OuterVolumeSpecName: "host-run-netns") pod "15117613-9fad-48c7-98c4-a2d84502ded9" (UID: "15117613-9fad-48c7-98c4-a2d84502ded9"). InnerVolumeSpecName "host-run-netns". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.631650 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-var-lib-openvswitch" (OuterVolumeSpecName: "var-lib-openvswitch") pod "15117613-9fad-48c7-98c4-a2d84502ded9" (UID: "15117613-9fad-48c7-98c4-a2d84502ded9"). InnerVolumeSpecName "var-lib-openvswitch". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.631915 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/15117613-9fad-48c7-98c4-a2d84502ded9-ovn-node-metrics-cert\") pod \"15117613-9fad-48c7-98c4-a2d84502ded9\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.631982 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/15117613-9fad-48c7-98c4-a2d84502ded9-ovnkube-config" (OuterVolumeSpecName: "ovnkube-config") pod "15117613-9fad-48c7-98c4-a2d84502ded9" (UID: "15117613-9fad-48c7-98c4-a2d84502ded9"). InnerVolumeSpecName "ovnkube-config". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.632172 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-systemd-units\") pod \"15117613-9fad-48c7-98c4-a2d84502ded9\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.632246 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-host-kubelet\") pod \"15117613-9fad-48c7-98c4-a2d84502ded9\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.632337 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-host-run-ovn-kubernetes\") pod \"15117613-9fad-48c7-98c4-a2d84502ded9\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.632420 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-host-slash\") pod \"15117613-9fad-48c7-98c4-a2d84502ded9\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.632502 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/15117613-9fad-48c7-98c4-a2d84502ded9-ovnkube-script-lib\") pod \"15117613-9fad-48c7-98c4-a2d84502ded9\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.632598 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/15117613-9fad-48c7-98c4-a2d84502ded9-env-overrides\") pod \"15117613-9fad-48c7-98c4-a2d84502ded9\" (UID: \"15117613-9fad-48c7-98c4-a2d84502ded9\") " Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.632827 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/cb755fc0-e934-4bef-93e9-427cbb2ee692-run-systemd\") pod \"ovnkube-node-lqrtr\" (UID: \"cb755fc0-e934-4bef-93e9-427cbb2ee692\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.632922 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/cb755fc0-e934-4bef-93e9-427cbb2ee692-host-cni-bin\") pod \"ovnkube-node-lqrtr\" (UID: \"cb755fc0-e934-4bef-93e9-427cbb2ee692\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.633003 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-smdbq\" (UniqueName: \"kubernetes.io/projected/cb755fc0-e934-4bef-93e9-427cbb2ee692-kube-api-access-smdbq\") pod \"ovnkube-node-lqrtr\" (UID: \"cb755fc0-e934-4bef-93e9-427cbb2ee692\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.633092 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/cb755fc0-e934-4bef-93e9-427cbb2ee692-log-socket\") pod \"ovnkube-node-lqrtr\" (UID: \"cb755fc0-e934-4bef-93e9-427cbb2ee692\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.633501 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-host-slash" (OuterVolumeSpecName: "host-slash") pod "15117613-9fad-48c7-98c4-a2d84502ded9" (UID: "15117613-9fad-48c7-98c4-a2d84502ded9"). InnerVolumeSpecName "host-slash". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.633531 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-systemd-units" (OuterVolumeSpecName: "systemd-units") pod "15117613-9fad-48c7-98c4-a2d84502ded9" (UID: "15117613-9fad-48c7-98c4-a2d84502ded9"). InnerVolumeSpecName "systemd-units". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.633548 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-host-kubelet" (OuterVolumeSpecName: "host-kubelet") pod "15117613-9fad-48c7-98c4-a2d84502ded9" (UID: "15117613-9fad-48c7-98c4-a2d84502ded9"). InnerVolumeSpecName "host-kubelet". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.633567 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-host-run-ovn-kubernetes" (OuterVolumeSpecName: "host-run-ovn-kubernetes") pod "15117613-9fad-48c7-98c4-a2d84502ded9" (UID: "15117613-9fad-48c7-98c4-a2d84502ded9"). InnerVolumeSpecName "host-run-ovn-kubernetes". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.634649 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/15117613-9fad-48c7-98c4-a2d84502ded9-env-overrides" (OuterVolumeSpecName: "env-overrides") pod "15117613-9fad-48c7-98c4-a2d84502ded9" (UID: "15117613-9fad-48c7-98c4-a2d84502ded9"). InnerVolumeSpecName "env-overrides". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.634923 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/cb755fc0-e934-4bef-93e9-427cbb2ee692-host-kubelet\") pod \"ovnkube-node-lqrtr\" (UID: \"cb755fc0-e934-4bef-93e9-427cbb2ee692\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.634967 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cb755fc0-e934-4bef-93e9-427cbb2ee692-var-lib-openvswitch\") pod \"ovnkube-node-lqrtr\" (UID: \"cb755fc0-e934-4bef-93e9-427cbb2ee692\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.635000 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/cb755fc0-e934-4bef-93e9-427cbb2ee692-run-ovn\") pod \"ovnkube-node-lqrtr\" (UID: \"cb755fc0-e934-4bef-93e9-427cbb2ee692\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.635027 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cb755fc0-e934-4bef-93e9-427cbb2ee692-run-openvswitch\") pod \"ovnkube-node-lqrtr\" (UID: \"cb755fc0-e934-4bef-93e9-427cbb2ee692\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.635082 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/cb755fc0-e934-4bef-93e9-427cbb2ee692-ovn-node-metrics-cert\") pod \"ovnkube-node-lqrtr\" (UID: \"cb755fc0-e934-4bef-93e9-427cbb2ee692\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.635132 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cb755fc0-e934-4bef-93e9-427cbb2ee692-etc-openvswitch\") pod \"ovnkube-node-lqrtr\" (UID: \"cb755fc0-e934-4bef-93e9-427cbb2ee692\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.635157 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/cb755fc0-e934-4bef-93e9-427cbb2ee692-host-run-netns\") pod \"ovnkube-node-lqrtr\" (UID: \"cb755fc0-e934-4bef-93e9-427cbb2ee692\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.635177 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/cb755fc0-e934-4bef-93e9-427cbb2ee692-host-slash\") pod \"ovnkube-node-lqrtr\" (UID: \"cb755fc0-e934-4bef-93e9-427cbb2ee692\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.635196 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/cb755fc0-e934-4bef-93e9-427cbb2ee692-systemd-units\") pod \"ovnkube-node-lqrtr\" (UID: \"cb755fc0-e934-4bef-93e9-427cbb2ee692\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.635223 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/cb755fc0-e934-4bef-93e9-427cbb2ee692-host-cni-netd\") pod \"ovnkube-node-lqrtr\" (UID: \"cb755fc0-e934-4bef-93e9-427cbb2ee692\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.635251 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/cb755fc0-e934-4bef-93e9-427cbb2ee692-host-run-ovn-kubernetes\") pod \"ovnkube-node-lqrtr\" (UID: \"cb755fc0-e934-4bef-93e9-427cbb2ee692\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.635360 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/cb755fc0-e934-4bef-93e9-427cbb2ee692-node-log\") pod \"ovnkube-node-lqrtr\" (UID: \"cb755fc0-e934-4bef-93e9-427cbb2ee692\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.635394 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/cb755fc0-e934-4bef-93e9-427cbb2ee692-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-lqrtr\" (UID: \"cb755fc0-e934-4bef-93e9-427cbb2ee692\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.635440 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/cb755fc0-e934-4bef-93e9-427cbb2ee692-ovnkube-script-lib\") pod \"ovnkube-node-lqrtr\" (UID: \"cb755fc0-e934-4bef-93e9-427cbb2ee692\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.635551 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/cb755fc0-e934-4bef-93e9-427cbb2ee692-env-overrides\") pod \"ovnkube-node-lqrtr\" (UID: \"cb755fc0-e934-4bef-93e9-427cbb2ee692\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.635592 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/cb755fc0-e934-4bef-93e9-427cbb2ee692-ovnkube-config\") pod \"ovnkube-node-lqrtr\" (UID: \"cb755fc0-e934-4bef-93e9-427cbb2ee692\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.635645 4948 reconciler_common.go:293] "Volume detached for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-var-lib-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.635657 4948 reconciler_common.go:293] "Volume detached for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-host-run-netns\") on node \"crc\" DevicePath \"\"" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.635666 4948 reconciler_common.go:293] "Volume detached for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-systemd-units\") on node \"crc\" DevicePath \"\"" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.635675 4948 reconciler_common.go:293] "Volume detached for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-host-kubelet\") on node \"crc\" DevicePath \"\"" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.635684 4948 reconciler_common.go:293] "Volume detached for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-host-run-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.635693 4948 reconciler_common.go:293] "Volume detached for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-host-slash\") on node \"crc\" DevicePath \"\"" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.635702 4948 reconciler_common.go:293] "Volume detached for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/15117613-9fad-48c7-98c4-a2d84502ded9-env-overrides\") on node \"crc\" DevicePath \"\"" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.635710 4948 reconciler_common.go:293] "Volume detached for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-host-cni-netd\") on node \"crc\" DevicePath \"\"" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.635717 4948 reconciler_common.go:293] "Volume detached for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-host-cni-bin\") on node \"crc\" DevicePath \"\"" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.635725 4948 reconciler_common.go:293] "Volume detached for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-log-socket\") on node \"crc\" DevicePath \"\"" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.635733 4948 reconciler_common.go:293] "Volume detached for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/15117613-9fad-48c7-98c4-a2d84502ded9-ovnkube-config\") on node \"crc\" DevicePath \"\"" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.635743 4948 reconciler_common.go:293] "Volume detached for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-host-var-lib-cni-networks-ovn-kubernetes\") on node \"crc\" DevicePath \"\"" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.635753 4948 reconciler_common.go:293] "Volume detached for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-run-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.635775 4948 reconciler_common.go:293] "Volume detached for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-etc-openvswitch\") on node \"crc\" DevicePath \"\"" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.635784 4948 reconciler_common.go:293] "Volume detached for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-node-log\") on node \"crc\" DevicePath \"\"" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.635796 4948 reconciler_common.go:293] "Volume detached for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-run-ovn\") on node \"crc\" DevicePath \"\"" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.636204 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/configmap/15117613-9fad-48c7-98c4-a2d84502ded9-ovnkube-script-lib" (OuterVolumeSpecName: "ovnkube-script-lib") pod "15117613-9fad-48c7-98c4-a2d84502ded9" (UID: "15117613-9fad-48c7-98c4-a2d84502ded9"). InnerVolumeSpecName "ovnkube-script-lib". PluginName "kubernetes.io/configmap", VolumeGidValue "" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.638253 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/15117613-9fad-48c7-98c4-a2d84502ded9-kube-api-access-dvwhv" (OuterVolumeSpecName: "kube-api-access-dvwhv") pod "15117613-9fad-48c7-98c4-a2d84502ded9" (UID: "15117613-9fad-48c7-98c4-a2d84502ded9"). InnerVolumeSpecName "kube-api-access-dvwhv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.640386 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/secret/15117613-9fad-48c7-98c4-a2d84502ded9-ovn-node-metrics-cert" (OuterVolumeSpecName: "ovn-node-metrics-cert") pod "15117613-9fad-48c7-98c4-a2d84502ded9" (UID: "15117613-9fad-48c7-98c4-a2d84502ded9"). InnerVolumeSpecName "ovn-node-metrics-cert". PluginName "kubernetes.io/secret", VolumeGidValue "" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.656925 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-run-systemd" (OuterVolumeSpecName: "run-systemd") pod "15117613-9fad-48c7-98c4-a2d84502ded9" (UID: "15117613-9fad-48c7-98c4-a2d84502ded9"). InnerVolumeSpecName "run-systemd". PluginName "kubernetes.io/host-path", VolumeGidValue "" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.736730 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/cb755fc0-e934-4bef-93e9-427cbb2ee692-node-log\") pod \"ovnkube-node-lqrtr\" (UID: \"cb755fc0-e934-4bef-93e9-427cbb2ee692\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.736802 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/cb755fc0-e934-4bef-93e9-427cbb2ee692-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-lqrtr\" (UID: \"cb755fc0-e934-4bef-93e9-427cbb2ee692\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.736846 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/cb755fc0-e934-4bef-93e9-427cbb2ee692-ovnkube-script-lib\") pod \"ovnkube-node-lqrtr\" (UID: \"cb755fc0-e934-4bef-93e9-427cbb2ee692\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.736900 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/cb755fc0-e934-4bef-93e9-427cbb2ee692-env-overrides\") pod \"ovnkube-node-lqrtr\" (UID: \"cb755fc0-e934-4bef-93e9-427cbb2ee692\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.736912 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"node-log\" (UniqueName: \"kubernetes.io/host-path/cb755fc0-e934-4bef-93e9-427cbb2ee692-node-log\") pod \"ovnkube-node-lqrtr\" (UID: \"cb755fc0-e934-4bef-93e9-427cbb2ee692\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.736931 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/cb755fc0-e934-4bef-93e9-427cbb2ee692-ovnkube-config\") pod \"ovnkube-node-lqrtr\" (UID: \"cb755fc0-e934-4bef-93e9-427cbb2ee692\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.737013 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/cb755fc0-e934-4bef-93e9-427cbb2ee692-run-systemd\") pod \"ovnkube-node-lqrtr\" (UID: \"cb755fc0-e934-4bef-93e9-427cbb2ee692\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.737047 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/cb755fc0-e934-4bef-93e9-427cbb2ee692-host-cni-bin\") pod \"ovnkube-node-lqrtr\" (UID: \"cb755fc0-e934-4bef-93e9-427cbb2ee692\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.737087 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-smdbq\" (UniqueName: \"kubernetes.io/projected/cb755fc0-e934-4bef-93e9-427cbb2ee692-kube-api-access-smdbq\") pod \"ovnkube-node-lqrtr\" (UID: \"cb755fc0-e934-4bef-93e9-427cbb2ee692\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.737127 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/cb755fc0-e934-4bef-93e9-427cbb2ee692-log-socket\") pod \"ovnkube-node-lqrtr\" (UID: \"cb755fc0-e934-4bef-93e9-427cbb2ee692\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.737162 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/cb755fc0-e934-4bef-93e9-427cbb2ee692-run-systemd\") pod \"ovnkube-node-lqrtr\" (UID: \"cb755fc0-e934-4bef-93e9-427cbb2ee692\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.737219 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/cb755fc0-e934-4bef-93e9-427cbb2ee692-host-kubelet\") pod \"ovnkube-node-lqrtr\" (UID: \"cb755fc0-e934-4bef-93e9-427cbb2ee692\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.737177 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-kubelet\" (UniqueName: \"kubernetes.io/host-path/cb755fc0-e934-4bef-93e9-427cbb2ee692-host-kubelet\") pod \"ovnkube-node-lqrtr\" (UID: \"cb755fc0-e934-4bef-93e9-427cbb2ee692\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.737216 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-bin\" (UniqueName: \"kubernetes.io/host-path/cb755fc0-e934-4bef-93e9-427cbb2ee692-host-cni-bin\") pod \"ovnkube-node-lqrtr\" (UID: \"cb755fc0-e934-4bef-93e9-427cbb2ee692\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.737295 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"log-socket\" (UniqueName: \"kubernetes.io/host-path/cb755fc0-e934-4bef-93e9-427cbb2ee692-log-socket\") pod \"ovnkube-node-lqrtr\" (UID: \"cb755fc0-e934-4bef-93e9-427cbb2ee692\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.737295 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cb755fc0-e934-4bef-93e9-427cbb2ee692-var-lib-openvswitch\") pod \"ovnkube-node-lqrtr\" (UID: \"cb755fc0-e934-4bef-93e9-427cbb2ee692\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.737424 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/cb755fc0-e934-4bef-93e9-427cbb2ee692-run-ovn\") pod \"ovnkube-node-lqrtr\" (UID: \"cb755fc0-e934-4bef-93e9-427cbb2ee692\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.737474 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cb755fc0-e934-4bef-93e9-427cbb2ee692-run-openvswitch\") pod \"ovnkube-node-lqrtr\" (UID: \"cb755fc0-e934-4bef-93e9-427cbb2ee692\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.737500 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"var-lib-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cb755fc0-e934-4bef-93e9-427cbb2ee692-var-lib-openvswitch\") pod \"ovnkube-node-lqrtr\" (UID: \"cb755fc0-e934-4bef-93e9-427cbb2ee692\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.737550 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-ovn\" (UniqueName: \"kubernetes.io/host-path/cb755fc0-e934-4bef-93e9-427cbb2ee692-run-ovn\") pod \"ovnkube-node-lqrtr\" (UID: \"cb755fc0-e934-4bef-93e9-427cbb2ee692\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.737616 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/cb755fc0-e934-4bef-93e9-427cbb2ee692-ovn-node-metrics-cert\") pod \"ovnkube-node-lqrtr\" (UID: \"cb755fc0-e934-4bef-93e9-427cbb2ee692\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.737657 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"run-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cb755fc0-e934-4bef-93e9-427cbb2ee692-run-openvswitch\") pod \"ovnkube-node-lqrtr\" (UID: \"cb755fc0-e934-4bef-93e9-427cbb2ee692\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.737692 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cb755fc0-e934-4bef-93e9-427cbb2ee692-etc-openvswitch\") pod \"ovnkube-node-lqrtr\" (UID: \"cb755fc0-e934-4bef-93e9-427cbb2ee692\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.737756 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"etc-openvswitch\" (UniqueName: \"kubernetes.io/host-path/cb755fc0-e934-4bef-93e9-427cbb2ee692-etc-openvswitch\") pod \"ovnkube-node-lqrtr\" (UID: \"cb755fc0-e934-4bef-93e9-427cbb2ee692\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.737624 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"env-overrides\" (UniqueName: \"kubernetes.io/configmap/cb755fc0-e934-4bef-93e9-427cbb2ee692-env-overrides\") pod \"ovnkube-node-lqrtr\" (UID: \"cb755fc0-e934-4bef-93e9-427cbb2ee692\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.737799 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/cb755fc0-e934-4bef-93e9-427cbb2ee692-host-run-netns\") pod \"ovnkube-node-lqrtr\" (UID: \"cb755fc0-e934-4bef-93e9-427cbb2ee692\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.737819 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/cb755fc0-e934-4bef-93e9-427cbb2ee692-ovnkube-script-lib\") pod \"ovnkube-node-lqrtr\" (UID: \"cb755fc0-e934-4bef-93e9-427cbb2ee692\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.737762 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-netns\" (UniqueName: \"kubernetes.io/host-path/cb755fc0-e934-4bef-93e9-427cbb2ee692-host-run-netns\") pod \"ovnkube-node-lqrtr\" (UID: \"cb755fc0-e934-4bef-93e9-427cbb2ee692\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.737856 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-var-lib-cni-networks-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/cb755fc0-e934-4bef-93e9-427cbb2ee692-host-var-lib-cni-networks-ovn-kubernetes\") pod \"ovnkube-node-lqrtr\" (UID: \"cb755fc0-e934-4bef-93e9-427cbb2ee692\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.737879 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/cb755fc0-e934-4bef-93e9-427cbb2ee692-host-slash\") pod \"ovnkube-node-lqrtr\" (UID: \"cb755fc0-e934-4bef-93e9-427cbb2ee692\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.737992 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/cb755fc0-e934-4bef-93e9-427cbb2ee692-systemd-units\") pod \"ovnkube-node-lqrtr\" (UID: \"cb755fc0-e934-4bef-93e9-427cbb2ee692\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.737910 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-slash\" (UniqueName: \"kubernetes.io/host-path/cb755fc0-e934-4bef-93e9-427cbb2ee692-host-slash\") pod \"ovnkube-node-lqrtr\" (UID: \"cb755fc0-e934-4bef-93e9-427cbb2ee692\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.738046 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/cb755fc0-e934-4bef-93e9-427cbb2ee692-host-run-ovn-kubernetes\") pod \"ovnkube-node-lqrtr\" (UID: \"cb755fc0-e934-4bef-93e9-427cbb2ee692\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.738085 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"systemd-units\" (UniqueName: \"kubernetes.io/host-path/cb755fc0-e934-4bef-93e9-427cbb2ee692-systemd-units\") pod \"ovnkube-node-lqrtr\" (UID: \"cb755fc0-e934-4bef-93e9-427cbb2ee692\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.738091 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/cb755fc0-e934-4bef-93e9-427cbb2ee692-host-cni-netd\") pod \"ovnkube-node-lqrtr\" (UID: \"cb755fc0-e934-4bef-93e9-427cbb2ee692\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.738144 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-cni-netd\" (UniqueName: \"kubernetes.io/host-path/cb755fc0-e934-4bef-93e9-427cbb2ee692-host-cni-netd\") pod \"ovnkube-node-lqrtr\" (UID: \"cb755fc0-e934-4bef-93e9-427cbb2ee692\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.738169 4948 reconciler_common.go:293] "Volume detached for volume \"run-systemd\" (UniqueName: \"kubernetes.io/host-path/15117613-9fad-48c7-98c4-a2d84502ded9-run-systemd\") on node \"crc\" DevicePath \"\"" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.738184 4948 reconciler_common.go:293] "Volume detached for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/15117613-9fad-48c7-98c4-a2d84502ded9-ovn-node-metrics-cert\") on node \"crc\" DevicePath \"\"" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.738190 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"host-run-ovn-kubernetes\" (UniqueName: \"kubernetes.io/host-path/cb755fc0-e934-4bef-93e9-427cbb2ee692-host-run-ovn-kubernetes\") pod \"ovnkube-node-lqrtr\" (UID: \"cb755fc0-e934-4bef-93e9-427cbb2ee692\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.738200 4948 reconciler_common.go:293] "Volume detached for volume \"ovnkube-script-lib\" (UniqueName: \"kubernetes.io/configmap/15117613-9fad-48c7-98c4-a2d84502ded9-ovnkube-script-lib\") on node \"crc\" DevicePath \"\"" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.738273 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dvwhv\" (UniqueName: \"kubernetes.io/projected/15117613-9fad-48c7-98c4-a2d84502ded9-kube-api-access-dvwhv\") on node \"crc\" DevicePath \"\"" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.738436 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovnkube-config\" (UniqueName: \"kubernetes.io/configmap/cb755fc0-e934-4bef-93e9-427cbb2ee692-ovnkube-config\") pod \"ovnkube-node-lqrtr\" (UID: \"cb755fc0-e934-4bef-93e9-427cbb2ee692\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.741901 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"ovn-node-metrics-cert\" (UniqueName: \"kubernetes.io/secret/cb755fc0-e934-4bef-93e9-427cbb2ee692-ovn-node-metrics-cert\") pod \"ovnkube-node-lqrtr\" (UID: \"cb755fc0-e934-4bef-93e9-427cbb2ee692\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.759356 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-smdbq\" (UniqueName: \"kubernetes.io/projected/cb755fc0-e934-4bef-93e9-427cbb2ee692-kube-api-access-smdbq\") pod \"ovnkube-node-lqrtr\" (UID: \"cb755fc0-e934-4bef-93e9-427cbb2ee692\") " pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.919197 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:18:38 crc kubenswrapper[4948]: W0312 00:18:38.952507 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podcb755fc0_e934_4bef_93e9_427cbb2ee692.slice/crio-cda20bfdbd21a11d7d53f5433e4b203039e1b3043a47420bb57af7aa2a30f9a4 WatchSource:0}: Error finding container cda20bfdbd21a11d7d53f5433e4b203039e1b3043a47420bb57af7aa2a30f9a4: Status 404 returned error can't find the container with id cda20bfdbd21a11d7d53f5433e4b203039e1b3043a47420bb57af7aa2a30f9a4 Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.952800 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-bwpqp_3b7580c2-5c23-4c67-807a-ea97a3df9398/kube-multus/2.log" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.953639 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-bwpqp_3b7580c2-5c23-4c67-807a-ea97a3df9398/kube-multus/1.log" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.953718 4948 generic.go:334] "Generic (PLEG): container finished" podID="3b7580c2-5c23-4c67-807a-ea97a3df9398" containerID="b283d9461d9f6b7694a42e106d5d1c94fa66752ee64832f229a2e6088e3fca81" exitCode=2 Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.953818 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-bwpqp" event={"ID":"3b7580c2-5c23-4c67-807a-ea97a3df9398","Type":"ContainerDied","Data":"b283d9461d9f6b7694a42e106d5d1c94fa66752ee64832f229a2e6088e3fca81"} Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.953943 4948 scope.go:117] "RemoveContainer" containerID="b28502ee0add02df45488b29f20d8be4102b0f3faaf8b5e26b5943a410163bf7" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.954697 4948 scope.go:117] "RemoveContainer" containerID="b283d9461d9f6b7694a42e106d5d1c94fa66752ee64832f229a2e6088e3fca81" Mar 12 00:18:38 crc kubenswrapper[4948]: E0312 00:18:38.955009 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-bwpqp_openshift-multus(3b7580c2-5c23-4c67-807a-ea97a3df9398)\"" pod="openshift-multus/multus-bwpqp" podUID="3b7580c2-5c23-4c67-807a-ea97a3df9398" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.962225 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f7mp5_15117613-9fad-48c7-98c4-a2d84502ded9/ovnkube-controller/3.log" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.967531 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f7mp5_15117613-9fad-48c7-98c4-a2d84502ded9/ovn-acl-logging/0.log" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.968361 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-ovn-kubernetes_ovnkube-node-f7mp5_15117613-9fad-48c7-98c4-a2d84502ded9/ovn-controller/0.log" Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.968898 4948 generic.go:334] "Generic (PLEG): container finished" podID="15117613-9fad-48c7-98c4-a2d84502ded9" containerID="d3de0f7f4aa39ee28267d3a0041033e89da7e51713f417910e09007110b7b9c2" exitCode=0 Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.968943 4948 generic.go:334] "Generic (PLEG): container finished" podID="15117613-9fad-48c7-98c4-a2d84502ded9" containerID="46fa6140f4f4bd32e298fa36f6d7af6890029f9bc3080d2b4d94be8307346de8" exitCode=0 Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.968960 4948 generic.go:334] "Generic (PLEG): container finished" podID="15117613-9fad-48c7-98c4-a2d84502ded9" containerID="2e481ace9e84b6aa66cbdda3ee8425c4120a665915d2c70ae4faf60bba6205e2" exitCode=0 Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.968977 4948 generic.go:334] "Generic (PLEG): container finished" podID="15117613-9fad-48c7-98c4-a2d84502ded9" containerID="c9eadefd0babc119f776d48c9d2c9e4bced58db685af9a8e97f25e6a9a2e3c89" exitCode=0 Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.968991 4948 generic.go:334] "Generic (PLEG): container finished" podID="15117613-9fad-48c7-98c4-a2d84502ded9" containerID="84af4cdd213d90636a889a8a8a3aaca3dee1e2677368efa0ffc768940c2cc994" exitCode=0 Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.969006 4948 generic.go:334] "Generic (PLEG): container finished" podID="15117613-9fad-48c7-98c4-a2d84502ded9" containerID="d4036d7131db00afd8eab8e093626b688297dc776b5af80c3c38b51da7124085" exitCode=0 Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.969021 4948 generic.go:334] "Generic (PLEG): container finished" podID="15117613-9fad-48c7-98c4-a2d84502ded9" containerID="ec9d5e77753fda70350353daa0967257dd4929dba4440a95495ab4c2a0659aef" exitCode=143 Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.969034 4948 generic.go:334] "Generic (PLEG): container finished" podID="15117613-9fad-48c7-98c4-a2d84502ded9" containerID="8cad943edb9a8d445b33a6e70d85b475c1ccba460a6056bccae94fe60fd6b73f" exitCode=143 Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.969078 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" event={"ID":"15117613-9fad-48c7-98c4-a2d84502ded9","Type":"ContainerDied","Data":"d3de0f7f4aa39ee28267d3a0041033e89da7e51713f417910e09007110b7b9c2"} Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.969121 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" event={"ID":"15117613-9fad-48c7-98c4-a2d84502ded9","Type":"ContainerDied","Data":"46fa6140f4f4bd32e298fa36f6d7af6890029f9bc3080d2b4d94be8307346de8"} Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.969146 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" event={"ID":"15117613-9fad-48c7-98c4-a2d84502ded9","Type":"ContainerDied","Data":"2e481ace9e84b6aa66cbdda3ee8425c4120a665915d2c70ae4faf60bba6205e2"} Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.969181 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" event={"ID":"15117613-9fad-48c7-98c4-a2d84502ded9","Type":"ContainerDied","Data":"c9eadefd0babc119f776d48c9d2c9e4bced58db685af9a8e97f25e6a9a2e3c89"} Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.969204 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" event={"ID":"15117613-9fad-48c7-98c4-a2d84502ded9","Type":"ContainerDied","Data":"84af4cdd213d90636a889a8a8a3aaca3dee1e2677368efa0ffc768940c2cc994"} Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.969226 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" event={"ID":"15117613-9fad-48c7-98c4-a2d84502ded9","Type":"ContainerDied","Data":"d4036d7131db00afd8eab8e093626b688297dc776b5af80c3c38b51da7124085"} Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.969254 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d3de0f7f4aa39ee28267d3a0041033e89da7e51713f417910e09007110b7b9c2"} Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.969271 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1d6a0a3fceb31bd31a6c978f0b3ac7ff768469ab3f311918e9dc72b424b00c27"} Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.969282 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"46fa6140f4f4bd32e298fa36f6d7af6890029f9bc3080d2b4d94be8307346de8"} Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.969294 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2e481ace9e84b6aa66cbdda3ee8425c4120a665915d2c70ae4faf60bba6205e2"} Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.969335 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c9eadefd0babc119f776d48c9d2c9e4bced58db685af9a8e97f25e6a9a2e3c89"} Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.969346 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"84af4cdd213d90636a889a8a8a3aaca3dee1e2677368efa0ffc768940c2cc994"} Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.969360 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d4036d7131db00afd8eab8e093626b688297dc776b5af80c3c38b51da7124085"} Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.969371 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ec9d5e77753fda70350353daa0967257dd4929dba4440a95495ab4c2a0659aef"} Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.969382 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8cad943edb9a8d445b33a6e70d85b475c1ccba460a6056bccae94fe60fd6b73f"} Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.969392 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a"} Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.969408 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" event={"ID":"15117613-9fad-48c7-98c4-a2d84502ded9","Type":"ContainerDied","Data":"ec9d5e77753fda70350353daa0967257dd4929dba4440a95495ab4c2a0659aef"} Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.969426 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d3de0f7f4aa39ee28267d3a0041033e89da7e51713f417910e09007110b7b9c2"} Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.969440 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1d6a0a3fceb31bd31a6c978f0b3ac7ff768469ab3f311918e9dc72b424b00c27"} Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.969451 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"46fa6140f4f4bd32e298fa36f6d7af6890029f9bc3080d2b4d94be8307346de8"} Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.969471 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2e481ace9e84b6aa66cbdda3ee8425c4120a665915d2c70ae4faf60bba6205e2"} Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.969484 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c9eadefd0babc119f776d48c9d2c9e4bced58db685af9a8e97f25e6a9a2e3c89"} Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.969498 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"84af4cdd213d90636a889a8a8a3aaca3dee1e2677368efa0ffc768940c2cc994"} Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.969511 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d4036d7131db00afd8eab8e093626b688297dc776b5af80c3c38b51da7124085"} Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.969525 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ec9d5e77753fda70350353daa0967257dd4929dba4440a95495ab4c2a0659aef"} Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.969538 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8cad943edb9a8d445b33a6e70d85b475c1ccba460a6056bccae94fe60fd6b73f"} Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.969551 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a"} Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.969572 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" event={"ID":"15117613-9fad-48c7-98c4-a2d84502ded9","Type":"ContainerDied","Data":"8cad943edb9a8d445b33a6e70d85b475c1ccba460a6056bccae94fe60fd6b73f"} Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.969597 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d3de0f7f4aa39ee28267d3a0041033e89da7e51713f417910e09007110b7b9c2"} Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.969613 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1d6a0a3fceb31bd31a6c978f0b3ac7ff768469ab3f311918e9dc72b424b00c27"} Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.969627 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"46fa6140f4f4bd32e298fa36f6d7af6890029f9bc3080d2b4d94be8307346de8"} Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.969641 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2e481ace9e84b6aa66cbdda3ee8425c4120a665915d2c70ae4faf60bba6205e2"} Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.969655 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c9eadefd0babc119f776d48c9d2c9e4bced58db685af9a8e97f25e6a9a2e3c89"} Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.969668 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"84af4cdd213d90636a889a8a8a3aaca3dee1e2677368efa0ffc768940c2cc994"} Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.969691 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d4036d7131db00afd8eab8e093626b688297dc776b5af80c3c38b51da7124085"} Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.969707 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ec9d5e77753fda70350353daa0967257dd4929dba4440a95495ab4c2a0659aef"} Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.969722 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8cad943edb9a8d445b33a6e70d85b475c1ccba460a6056bccae94fe60fd6b73f"} Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.969736 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a"} Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.969752 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" event={"ID":"15117613-9fad-48c7-98c4-a2d84502ded9","Type":"ContainerDied","Data":"2b5e9ab6e8d41452dcec8681682d1d5e7bd916598cb0e81980f0fc3ac545e902"} Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.969771 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d3de0f7f4aa39ee28267d3a0041033e89da7e51713f417910e09007110b7b9c2"} Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.969784 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"1d6a0a3fceb31bd31a6c978f0b3ac7ff768469ab3f311918e9dc72b424b00c27"} Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.969798 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"46fa6140f4f4bd32e298fa36f6d7af6890029f9bc3080d2b4d94be8307346de8"} Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.969811 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"2e481ace9e84b6aa66cbdda3ee8425c4120a665915d2c70ae4faf60bba6205e2"} Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.969825 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c9eadefd0babc119f776d48c9d2c9e4bced58db685af9a8e97f25e6a9a2e3c89"} Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.969840 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"84af4cdd213d90636a889a8a8a3aaca3dee1e2677368efa0ffc768940c2cc994"} Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.969854 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"d4036d7131db00afd8eab8e093626b688297dc776b5af80c3c38b51da7124085"} Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.969868 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"ec9d5e77753fda70350353daa0967257dd4929dba4440a95495ab4c2a0659aef"} Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.969881 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"8cad943edb9a8d445b33a6e70d85b475c1ccba460a6056bccae94fe60fd6b73f"} Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.969903 4948 pod_container_deletor.go:114] "Failed to issue the request to remove container" containerID={"Type":"cri-o","ID":"c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a"} Mar 12 00:18:38 crc kubenswrapper[4948]: I0312 00:18:38.970719 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-ovn-kubernetes/ovnkube-node-f7mp5" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.004965 4948 scope.go:117] "RemoveContainer" containerID="d3de0f7f4aa39ee28267d3a0041033e89da7e51713f417910e09007110b7b9c2" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.029831 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-f7mp5"] Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.042436 4948 scope.go:117] "RemoveContainer" containerID="1d6a0a3fceb31bd31a6c978f0b3ac7ff768469ab3f311918e9dc72b424b00c27" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.042990 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-ovn-kubernetes/ovnkube-node-f7mp5"] Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.070138 4948 scope.go:117] "RemoveContainer" containerID="46fa6140f4f4bd32e298fa36f6d7af6890029f9bc3080d2b4d94be8307346de8" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.101941 4948 scope.go:117] "RemoveContainer" containerID="2e481ace9e84b6aa66cbdda3ee8425c4120a665915d2c70ae4faf60bba6205e2" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.129699 4948 scope.go:117] "RemoveContainer" containerID="c9eadefd0babc119f776d48c9d2c9e4bced58db685af9a8e97f25e6a9a2e3c89" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.150533 4948 scope.go:117] "RemoveContainer" containerID="84af4cdd213d90636a889a8a8a3aaca3dee1e2677368efa0ffc768940c2cc994" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.221605 4948 scope.go:117] "RemoveContainer" containerID="d4036d7131db00afd8eab8e093626b688297dc776b5af80c3c38b51da7124085" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.245118 4948 scope.go:117] "RemoveContainer" containerID="ec9d5e77753fda70350353daa0967257dd4929dba4440a95495ab4c2a0659aef" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.267517 4948 scope.go:117] "RemoveContainer" containerID="8cad943edb9a8d445b33a6e70d85b475c1ccba460a6056bccae94fe60fd6b73f" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.325617 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="15117613-9fad-48c7-98c4-a2d84502ded9" path="/var/lib/kubelet/pods/15117613-9fad-48c7-98c4-a2d84502ded9/volumes" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.327510 4948 scope.go:117] "RemoveContainer" containerID="c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.346082 4948 scope.go:117] "RemoveContainer" containerID="d3de0f7f4aa39ee28267d3a0041033e89da7e51713f417910e09007110b7b9c2" Mar 12 00:18:39 crc kubenswrapper[4948]: E0312 00:18:39.346859 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d3de0f7f4aa39ee28267d3a0041033e89da7e51713f417910e09007110b7b9c2\": container with ID starting with d3de0f7f4aa39ee28267d3a0041033e89da7e51713f417910e09007110b7b9c2 not found: ID does not exist" containerID="d3de0f7f4aa39ee28267d3a0041033e89da7e51713f417910e09007110b7b9c2" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.346927 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3de0f7f4aa39ee28267d3a0041033e89da7e51713f417910e09007110b7b9c2"} err="failed to get container status \"d3de0f7f4aa39ee28267d3a0041033e89da7e51713f417910e09007110b7b9c2\": rpc error: code = NotFound desc = could not find container \"d3de0f7f4aa39ee28267d3a0041033e89da7e51713f417910e09007110b7b9c2\": container with ID starting with d3de0f7f4aa39ee28267d3a0041033e89da7e51713f417910e09007110b7b9c2 not found: ID does not exist" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.346970 4948 scope.go:117] "RemoveContainer" containerID="1d6a0a3fceb31bd31a6c978f0b3ac7ff768469ab3f311918e9dc72b424b00c27" Mar 12 00:18:39 crc kubenswrapper[4948]: E0312 00:18:39.347698 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"1d6a0a3fceb31bd31a6c978f0b3ac7ff768469ab3f311918e9dc72b424b00c27\": container with ID starting with 1d6a0a3fceb31bd31a6c978f0b3ac7ff768469ab3f311918e9dc72b424b00c27 not found: ID does not exist" containerID="1d6a0a3fceb31bd31a6c978f0b3ac7ff768469ab3f311918e9dc72b424b00c27" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.347752 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d6a0a3fceb31bd31a6c978f0b3ac7ff768469ab3f311918e9dc72b424b00c27"} err="failed to get container status \"1d6a0a3fceb31bd31a6c978f0b3ac7ff768469ab3f311918e9dc72b424b00c27\": rpc error: code = NotFound desc = could not find container \"1d6a0a3fceb31bd31a6c978f0b3ac7ff768469ab3f311918e9dc72b424b00c27\": container with ID starting with 1d6a0a3fceb31bd31a6c978f0b3ac7ff768469ab3f311918e9dc72b424b00c27 not found: ID does not exist" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.347784 4948 scope.go:117] "RemoveContainer" containerID="46fa6140f4f4bd32e298fa36f6d7af6890029f9bc3080d2b4d94be8307346de8" Mar 12 00:18:39 crc kubenswrapper[4948]: E0312 00:18:39.348431 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"46fa6140f4f4bd32e298fa36f6d7af6890029f9bc3080d2b4d94be8307346de8\": container with ID starting with 46fa6140f4f4bd32e298fa36f6d7af6890029f9bc3080d2b4d94be8307346de8 not found: ID does not exist" containerID="46fa6140f4f4bd32e298fa36f6d7af6890029f9bc3080d2b4d94be8307346de8" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.348481 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46fa6140f4f4bd32e298fa36f6d7af6890029f9bc3080d2b4d94be8307346de8"} err="failed to get container status \"46fa6140f4f4bd32e298fa36f6d7af6890029f9bc3080d2b4d94be8307346de8\": rpc error: code = NotFound desc = could not find container \"46fa6140f4f4bd32e298fa36f6d7af6890029f9bc3080d2b4d94be8307346de8\": container with ID starting with 46fa6140f4f4bd32e298fa36f6d7af6890029f9bc3080d2b4d94be8307346de8 not found: ID does not exist" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.348509 4948 scope.go:117] "RemoveContainer" containerID="2e481ace9e84b6aa66cbdda3ee8425c4120a665915d2c70ae4faf60bba6205e2" Mar 12 00:18:39 crc kubenswrapper[4948]: E0312 00:18:39.350265 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"2e481ace9e84b6aa66cbdda3ee8425c4120a665915d2c70ae4faf60bba6205e2\": container with ID starting with 2e481ace9e84b6aa66cbdda3ee8425c4120a665915d2c70ae4faf60bba6205e2 not found: ID does not exist" containerID="2e481ace9e84b6aa66cbdda3ee8425c4120a665915d2c70ae4faf60bba6205e2" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.350350 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e481ace9e84b6aa66cbdda3ee8425c4120a665915d2c70ae4faf60bba6205e2"} err="failed to get container status \"2e481ace9e84b6aa66cbdda3ee8425c4120a665915d2c70ae4faf60bba6205e2\": rpc error: code = NotFound desc = could not find container \"2e481ace9e84b6aa66cbdda3ee8425c4120a665915d2c70ae4faf60bba6205e2\": container with ID starting with 2e481ace9e84b6aa66cbdda3ee8425c4120a665915d2c70ae4faf60bba6205e2 not found: ID does not exist" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.350388 4948 scope.go:117] "RemoveContainer" containerID="c9eadefd0babc119f776d48c9d2c9e4bced58db685af9a8e97f25e6a9a2e3c89" Mar 12 00:18:39 crc kubenswrapper[4948]: E0312 00:18:39.351712 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c9eadefd0babc119f776d48c9d2c9e4bced58db685af9a8e97f25e6a9a2e3c89\": container with ID starting with c9eadefd0babc119f776d48c9d2c9e4bced58db685af9a8e97f25e6a9a2e3c89 not found: ID does not exist" containerID="c9eadefd0babc119f776d48c9d2c9e4bced58db685af9a8e97f25e6a9a2e3c89" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.351959 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9eadefd0babc119f776d48c9d2c9e4bced58db685af9a8e97f25e6a9a2e3c89"} err="failed to get container status \"c9eadefd0babc119f776d48c9d2c9e4bced58db685af9a8e97f25e6a9a2e3c89\": rpc error: code = NotFound desc = could not find container \"c9eadefd0babc119f776d48c9d2c9e4bced58db685af9a8e97f25e6a9a2e3c89\": container with ID starting with c9eadefd0babc119f776d48c9d2c9e4bced58db685af9a8e97f25e6a9a2e3c89 not found: ID does not exist" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.352056 4948 scope.go:117] "RemoveContainer" containerID="84af4cdd213d90636a889a8a8a3aaca3dee1e2677368efa0ffc768940c2cc994" Mar 12 00:18:39 crc kubenswrapper[4948]: E0312 00:18:39.352855 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"84af4cdd213d90636a889a8a8a3aaca3dee1e2677368efa0ffc768940c2cc994\": container with ID starting with 84af4cdd213d90636a889a8a8a3aaca3dee1e2677368efa0ffc768940c2cc994 not found: ID does not exist" containerID="84af4cdd213d90636a889a8a8a3aaca3dee1e2677368efa0ffc768940c2cc994" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.352920 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84af4cdd213d90636a889a8a8a3aaca3dee1e2677368efa0ffc768940c2cc994"} err="failed to get container status \"84af4cdd213d90636a889a8a8a3aaca3dee1e2677368efa0ffc768940c2cc994\": rpc error: code = NotFound desc = could not find container \"84af4cdd213d90636a889a8a8a3aaca3dee1e2677368efa0ffc768940c2cc994\": container with ID starting with 84af4cdd213d90636a889a8a8a3aaca3dee1e2677368efa0ffc768940c2cc994 not found: ID does not exist" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.352959 4948 scope.go:117] "RemoveContainer" containerID="d4036d7131db00afd8eab8e093626b688297dc776b5af80c3c38b51da7124085" Mar 12 00:18:39 crc kubenswrapper[4948]: E0312 00:18:39.353525 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d4036d7131db00afd8eab8e093626b688297dc776b5af80c3c38b51da7124085\": container with ID starting with d4036d7131db00afd8eab8e093626b688297dc776b5af80c3c38b51da7124085 not found: ID does not exist" containerID="d4036d7131db00afd8eab8e093626b688297dc776b5af80c3c38b51da7124085" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.353561 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4036d7131db00afd8eab8e093626b688297dc776b5af80c3c38b51da7124085"} err="failed to get container status \"d4036d7131db00afd8eab8e093626b688297dc776b5af80c3c38b51da7124085\": rpc error: code = NotFound desc = could not find container \"d4036d7131db00afd8eab8e093626b688297dc776b5af80c3c38b51da7124085\": container with ID starting with d4036d7131db00afd8eab8e093626b688297dc776b5af80c3c38b51da7124085 not found: ID does not exist" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.353583 4948 scope.go:117] "RemoveContainer" containerID="ec9d5e77753fda70350353daa0967257dd4929dba4440a95495ab4c2a0659aef" Mar 12 00:18:39 crc kubenswrapper[4948]: E0312 00:18:39.354089 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"ec9d5e77753fda70350353daa0967257dd4929dba4440a95495ab4c2a0659aef\": container with ID starting with ec9d5e77753fda70350353daa0967257dd4929dba4440a95495ab4c2a0659aef not found: ID does not exist" containerID="ec9d5e77753fda70350353daa0967257dd4929dba4440a95495ab4c2a0659aef" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.354116 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec9d5e77753fda70350353daa0967257dd4929dba4440a95495ab4c2a0659aef"} err="failed to get container status \"ec9d5e77753fda70350353daa0967257dd4929dba4440a95495ab4c2a0659aef\": rpc error: code = NotFound desc = could not find container \"ec9d5e77753fda70350353daa0967257dd4929dba4440a95495ab4c2a0659aef\": container with ID starting with ec9d5e77753fda70350353daa0967257dd4929dba4440a95495ab4c2a0659aef not found: ID does not exist" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.354136 4948 scope.go:117] "RemoveContainer" containerID="8cad943edb9a8d445b33a6e70d85b475c1ccba460a6056bccae94fe60fd6b73f" Mar 12 00:18:39 crc kubenswrapper[4948]: E0312 00:18:39.354866 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"8cad943edb9a8d445b33a6e70d85b475c1ccba460a6056bccae94fe60fd6b73f\": container with ID starting with 8cad943edb9a8d445b33a6e70d85b475c1ccba460a6056bccae94fe60fd6b73f not found: ID does not exist" containerID="8cad943edb9a8d445b33a6e70d85b475c1ccba460a6056bccae94fe60fd6b73f" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.354899 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8cad943edb9a8d445b33a6e70d85b475c1ccba460a6056bccae94fe60fd6b73f"} err="failed to get container status \"8cad943edb9a8d445b33a6e70d85b475c1ccba460a6056bccae94fe60fd6b73f\": rpc error: code = NotFound desc = could not find container \"8cad943edb9a8d445b33a6e70d85b475c1ccba460a6056bccae94fe60fd6b73f\": container with ID starting with 8cad943edb9a8d445b33a6e70d85b475c1ccba460a6056bccae94fe60fd6b73f not found: ID does not exist" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.354918 4948 scope.go:117] "RemoveContainer" containerID="c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a" Mar 12 00:18:39 crc kubenswrapper[4948]: E0312 00:18:39.355730 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\": container with ID starting with c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a not found: ID does not exist" containerID="c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.355815 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a"} err="failed to get container status \"c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\": rpc error: code = NotFound desc = could not find container \"c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\": container with ID starting with c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a not found: ID does not exist" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.355846 4948 scope.go:117] "RemoveContainer" containerID="d3de0f7f4aa39ee28267d3a0041033e89da7e51713f417910e09007110b7b9c2" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.356360 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3de0f7f4aa39ee28267d3a0041033e89da7e51713f417910e09007110b7b9c2"} err="failed to get container status \"d3de0f7f4aa39ee28267d3a0041033e89da7e51713f417910e09007110b7b9c2\": rpc error: code = NotFound desc = could not find container \"d3de0f7f4aa39ee28267d3a0041033e89da7e51713f417910e09007110b7b9c2\": container with ID starting with d3de0f7f4aa39ee28267d3a0041033e89da7e51713f417910e09007110b7b9c2 not found: ID does not exist" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.356416 4948 scope.go:117] "RemoveContainer" containerID="1d6a0a3fceb31bd31a6c978f0b3ac7ff768469ab3f311918e9dc72b424b00c27" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.356820 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d6a0a3fceb31bd31a6c978f0b3ac7ff768469ab3f311918e9dc72b424b00c27"} err="failed to get container status \"1d6a0a3fceb31bd31a6c978f0b3ac7ff768469ab3f311918e9dc72b424b00c27\": rpc error: code = NotFound desc = could not find container \"1d6a0a3fceb31bd31a6c978f0b3ac7ff768469ab3f311918e9dc72b424b00c27\": container with ID starting with 1d6a0a3fceb31bd31a6c978f0b3ac7ff768469ab3f311918e9dc72b424b00c27 not found: ID does not exist" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.356845 4948 scope.go:117] "RemoveContainer" containerID="46fa6140f4f4bd32e298fa36f6d7af6890029f9bc3080d2b4d94be8307346de8" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.357525 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46fa6140f4f4bd32e298fa36f6d7af6890029f9bc3080d2b4d94be8307346de8"} err="failed to get container status \"46fa6140f4f4bd32e298fa36f6d7af6890029f9bc3080d2b4d94be8307346de8\": rpc error: code = NotFound desc = could not find container \"46fa6140f4f4bd32e298fa36f6d7af6890029f9bc3080d2b4d94be8307346de8\": container with ID starting with 46fa6140f4f4bd32e298fa36f6d7af6890029f9bc3080d2b4d94be8307346de8 not found: ID does not exist" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.357600 4948 scope.go:117] "RemoveContainer" containerID="2e481ace9e84b6aa66cbdda3ee8425c4120a665915d2c70ae4faf60bba6205e2" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.358171 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e481ace9e84b6aa66cbdda3ee8425c4120a665915d2c70ae4faf60bba6205e2"} err="failed to get container status \"2e481ace9e84b6aa66cbdda3ee8425c4120a665915d2c70ae4faf60bba6205e2\": rpc error: code = NotFound desc = could not find container \"2e481ace9e84b6aa66cbdda3ee8425c4120a665915d2c70ae4faf60bba6205e2\": container with ID starting with 2e481ace9e84b6aa66cbdda3ee8425c4120a665915d2c70ae4faf60bba6205e2 not found: ID does not exist" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.358199 4948 scope.go:117] "RemoveContainer" containerID="c9eadefd0babc119f776d48c9d2c9e4bced58db685af9a8e97f25e6a9a2e3c89" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.358740 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9eadefd0babc119f776d48c9d2c9e4bced58db685af9a8e97f25e6a9a2e3c89"} err="failed to get container status \"c9eadefd0babc119f776d48c9d2c9e4bced58db685af9a8e97f25e6a9a2e3c89\": rpc error: code = NotFound desc = could not find container \"c9eadefd0babc119f776d48c9d2c9e4bced58db685af9a8e97f25e6a9a2e3c89\": container with ID starting with c9eadefd0babc119f776d48c9d2c9e4bced58db685af9a8e97f25e6a9a2e3c89 not found: ID does not exist" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.358791 4948 scope.go:117] "RemoveContainer" containerID="84af4cdd213d90636a889a8a8a3aaca3dee1e2677368efa0ffc768940c2cc994" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.359206 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84af4cdd213d90636a889a8a8a3aaca3dee1e2677368efa0ffc768940c2cc994"} err="failed to get container status \"84af4cdd213d90636a889a8a8a3aaca3dee1e2677368efa0ffc768940c2cc994\": rpc error: code = NotFound desc = could not find container \"84af4cdd213d90636a889a8a8a3aaca3dee1e2677368efa0ffc768940c2cc994\": container with ID starting with 84af4cdd213d90636a889a8a8a3aaca3dee1e2677368efa0ffc768940c2cc994 not found: ID does not exist" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.359232 4948 scope.go:117] "RemoveContainer" containerID="d4036d7131db00afd8eab8e093626b688297dc776b5af80c3c38b51da7124085" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.359834 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4036d7131db00afd8eab8e093626b688297dc776b5af80c3c38b51da7124085"} err="failed to get container status \"d4036d7131db00afd8eab8e093626b688297dc776b5af80c3c38b51da7124085\": rpc error: code = NotFound desc = could not find container \"d4036d7131db00afd8eab8e093626b688297dc776b5af80c3c38b51da7124085\": container with ID starting with d4036d7131db00afd8eab8e093626b688297dc776b5af80c3c38b51da7124085 not found: ID does not exist" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.359865 4948 scope.go:117] "RemoveContainer" containerID="ec9d5e77753fda70350353daa0967257dd4929dba4440a95495ab4c2a0659aef" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.360662 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec9d5e77753fda70350353daa0967257dd4929dba4440a95495ab4c2a0659aef"} err="failed to get container status \"ec9d5e77753fda70350353daa0967257dd4929dba4440a95495ab4c2a0659aef\": rpc error: code = NotFound desc = could not find container \"ec9d5e77753fda70350353daa0967257dd4929dba4440a95495ab4c2a0659aef\": container with ID starting with ec9d5e77753fda70350353daa0967257dd4929dba4440a95495ab4c2a0659aef not found: ID does not exist" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.360720 4948 scope.go:117] "RemoveContainer" containerID="8cad943edb9a8d445b33a6e70d85b475c1ccba460a6056bccae94fe60fd6b73f" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.361265 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8cad943edb9a8d445b33a6e70d85b475c1ccba460a6056bccae94fe60fd6b73f"} err="failed to get container status \"8cad943edb9a8d445b33a6e70d85b475c1ccba460a6056bccae94fe60fd6b73f\": rpc error: code = NotFound desc = could not find container \"8cad943edb9a8d445b33a6e70d85b475c1ccba460a6056bccae94fe60fd6b73f\": container with ID starting with 8cad943edb9a8d445b33a6e70d85b475c1ccba460a6056bccae94fe60fd6b73f not found: ID does not exist" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.361355 4948 scope.go:117] "RemoveContainer" containerID="c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.361834 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a"} err="failed to get container status \"c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\": rpc error: code = NotFound desc = could not find container \"c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\": container with ID starting with c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a not found: ID does not exist" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.361862 4948 scope.go:117] "RemoveContainer" containerID="d3de0f7f4aa39ee28267d3a0041033e89da7e51713f417910e09007110b7b9c2" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.362348 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3de0f7f4aa39ee28267d3a0041033e89da7e51713f417910e09007110b7b9c2"} err="failed to get container status \"d3de0f7f4aa39ee28267d3a0041033e89da7e51713f417910e09007110b7b9c2\": rpc error: code = NotFound desc = could not find container \"d3de0f7f4aa39ee28267d3a0041033e89da7e51713f417910e09007110b7b9c2\": container with ID starting with d3de0f7f4aa39ee28267d3a0041033e89da7e51713f417910e09007110b7b9c2 not found: ID does not exist" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.362399 4948 scope.go:117] "RemoveContainer" containerID="1d6a0a3fceb31bd31a6c978f0b3ac7ff768469ab3f311918e9dc72b424b00c27" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.362806 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d6a0a3fceb31bd31a6c978f0b3ac7ff768469ab3f311918e9dc72b424b00c27"} err="failed to get container status \"1d6a0a3fceb31bd31a6c978f0b3ac7ff768469ab3f311918e9dc72b424b00c27\": rpc error: code = NotFound desc = could not find container \"1d6a0a3fceb31bd31a6c978f0b3ac7ff768469ab3f311918e9dc72b424b00c27\": container with ID starting with 1d6a0a3fceb31bd31a6c978f0b3ac7ff768469ab3f311918e9dc72b424b00c27 not found: ID does not exist" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.362832 4948 scope.go:117] "RemoveContainer" containerID="46fa6140f4f4bd32e298fa36f6d7af6890029f9bc3080d2b4d94be8307346de8" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.363419 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46fa6140f4f4bd32e298fa36f6d7af6890029f9bc3080d2b4d94be8307346de8"} err="failed to get container status \"46fa6140f4f4bd32e298fa36f6d7af6890029f9bc3080d2b4d94be8307346de8\": rpc error: code = NotFound desc = could not find container \"46fa6140f4f4bd32e298fa36f6d7af6890029f9bc3080d2b4d94be8307346de8\": container with ID starting with 46fa6140f4f4bd32e298fa36f6d7af6890029f9bc3080d2b4d94be8307346de8 not found: ID does not exist" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.363534 4948 scope.go:117] "RemoveContainer" containerID="2e481ace9e84b6aa66cbdda3ee8425c4120a665915d2c70ae4faf60bba6205e2" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.364100 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e481ace9e84b6aa66cbdda3ee8425c4120a665915d2c70ae4faf60bba6205e2"} err="failed to get container status \"2e481ace9e84b6aa66cbdda3ee8425c4120a665915d2c70ae4faf60bba6205e2\": rpc error: code = NotFound desc = could not find container \"2e481ace9e84b6aa66cbdda3ee8425c4120a665915d2c70ae4faf60bba6205e2\": container with ID starting with 2e481ace9e84b6aa66cbdda3ee8425c4120a665915d2c70ae4faf60bba6205e2 not found: ID does not exist" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.364158 4948 scope.go:117] "RemoveContainer" containerID="c9eadefd0babc119f776d48c9d2c9e4bced58db685af9a8e97f25e6a9a2e3c89" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.364662 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9eadefd0babc119f776d48c9d2c9e4bced58db685af9a8e97f25e6a9a2e3c89"} err="failed to get container status \"c9eadefd0babc119f776d48c9d2c9e4bced58db685af9a8e97f25e6a9a2e3c89\": rpc error: code = NotFound desc = could not find container \"c9eadefd0babc119f776d48c9d2c9e4bced58db685af9a8e97f25e6a9a2e3c89\": container with ID starting with c9eadefd0babc119f776d48c9d2c9e4bced58db685af9a8e97f25e6a9a2e3c89 not found: ID does not exist" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.364688 4948 scope.go:117] "RemoveContainer" containerID="84af4cdd213d90636a889a8a8a3aaca3dee1e2677368efa0ffc768940c2cc994" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.365079 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84af4cdd213d90636a889a8a8a3aaca3dee1e2677368efa0ffc768940c2cc994"} err="failed to get container status \"84af4cdd213d90636a889a8a8a3aaca3dee1e2677368efa0ffc768940c2cc994\": rpc error: code = NotFound desc = could not find container \"84af4cdd213d90636a889a8a8a3aaca3dee1e2677368efa0ffc768940c2cc994\": container with ID starting with 84af4cdd213d90636a889a8a8a3aaca3dee1e2677368efa0ffc768940c2cc994 not found: ID does not exist" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.365101 4948 scope.go:117] "RemoveContainer" containerID="d4036d7131db00afd8eab8e093626b688297dc776b5af80c3c38b51da7124085" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.365694 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4036d7131db00afd8eab8e093626b688297dc776b5af80c3c38b51da7124085"} err="failed to get container status \"d4036d7131db00afd8eab8e093626b688297dc776b5af80c3c38b51da7124085\": rpc error: code = NotFound desc = could not find container \"d4036d7131db00afd8eab8e093626b688297dc776b5af80c3c38b51da7124085\": container with ID starting with d4036d7131db00afd8eab8e093626b688297dc776b5af80c3c38b51da7124085 not found: ID does not exist" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.365753 4948 scope.go:117] "RemoveContainer" containerID="ec9d5e77753fda70350353daa0967257dd4929dba4440a95495ab4c2a0659aef" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.366250 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec9d5e77753fda70350353daa0967257dd4929dba4440a95495ab4c2a0659aef"} err="failed to get container status \"ec9d5e77753fda70350353daa0967257dd4929dba4440a95495ab4c2a0659aef\": rpc error: code = NotFound desc = could not find container \"ec9d5e77753fda70350353daa0967257dd4929dba4440a95495ab4c2a0659aef\": container with ID starting with ec9d5e77753fda70350353daa0967257dd4929dba4440a95495ab4c2a0659aef not found: ID does not exist" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.366579 4948 scope.go:117] "RemoveContainer" containerID="8cad943edb9a8d445b33a6e70d85b475c1ccba460a6056bccae94fe60fd6b73f" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.366958 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8cad943edb9a8d445b33a6e70d85b475c1ccba460a6056bccae94fe60fd6b73f"} err="failed to get container status \"8cad943edb9a8d445b33a6e70d85b475c1ccba460a6056bccae94fe60fd6b73f\": rpc error: code = NotFound desc = could not find container \"8cad943edb9a8d445b33a6e70d85b475c1ccba460a6056bccae94fe60fd6b73f\": container with ID starting with 8cad943edb9a8d445b33a6e70d85b475c1ccba460a6056bccae94fe60fd6b73f not found: ID does not exist" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.366982 4948 scope.go:117] "RemoveContainer" containerID="c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.367466 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a"} err="failed to get container status \"c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\": rpc error: code = NotFound desc = could not find container \"c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\": container with ID starting with c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a not found: ID does not exist" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.367521 4948 scope.go:117] "RemoveContainer" containerID="d3de0f7f4aa39ee28267d3a0041033e89da7e51713f417910e09007110b7b9c2" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.368000 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d3de0f7f4aa39ee28267d3a0041033e89da7e51713f417910e09007110b7b9c2"} err="failed to get container status \"d3de0f7f4aa39ee28267d3a0041033e89da7e51713f417910e09007110b7b9c2\": rpc error: code = NotFound desc = could not find container \"d3de0f7f4aa39ee28267d3a0041033e89da7e51713f417910e09007110b7b9c2\": container with ID starting with d3de0f7f4aa39ee28267d3a0041033e89da7e51713f417910e09007110b7b9c2 not found: ID does not exist" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.368046 4948 scope.go:117] "RemoveContainer" containerID="1d6a0a3fceb31bd31a6c978f0b3ac7ff768469ab3f311918e9dc72b424b00c27" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.368517 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"1d6a0a3fceb31bd31a6c978f0b3ac7ff768469ab3f311918e9dc72b424b00c27"} err="failed to get container status \"1d6a0a3fceb31bd31a6c978f0b3ac7ff768469ab3f311918e9dc72b424b00c27\": rpc error: code = NotFound desc = could not find container \"1d6a0a3fceb31bd31a6c978f0b3ac7ff768469ab3f311918e9dc72b424b00c27\": container with ID starting with 1d6a0a3fceb31bd31a6c978f0b3ac7ff768469ab3f311918e9dc72b424b00c27 not found: ID does not exist" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.368540 4948 scope.go:117] "RemoveContainer" containerID="46fa6140f4f4bd32e298fa36f6d7af6890029f9bc3080d2b4d94be8307346de8" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.369427 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"46fa6140f4f4bd32e298fa36f6d7af6890029f9bc3080d2b4d94be8307346de8"} err="failed to get container status \"46fa6140f4f4bd32e298fa36f6d7af6890029f9bc3080d2b4d94be8307346de8\": rpc error: code = NotFound desc = could not find container \"46fa6140f4f4bd32e298fa36f6d7af6890029f9bc3080d2b4d94be8307346de8\": container with ID starting with 46fa6140f4f4bd32e298fa36f6d7af6890029f9bc3080d2b4d94be8307346de8 not found: ID does not exist" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.369472 4948 scope.go:117] "RemoveContainer" containerID="2e481ace9e84b6aa66cbdda3ee8425c4120a665915d2c70ae4faf60bba6205e2" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.369993 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"2e481ace9e84b6aa66cbdda3ee8425c4120a665915d2c70ae4faf60bba6205e2"} err="failed to get container status \"2e481ace9e84b6aa66cbdda3ee8425c4120a665915d2c70ae4faf60bba6205e2\": rpc error: code = NotFound desc = could not find container \"2e481ace9e84b6aa66cbdda3ee8425c4120a665915d2c70ae4faf60bba6205e2\": container with ID starting with 2e481ace9e84b6aa66cbdda3ee8425c4120a665915d2c70ae4faf60bba6205e2 not found: ID does not exist" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.370043 4948 scope.go:117] "RemoveContainer" containerID="c9eadefd0babc119f776d48c9d2c9e4bced58db685af9a8e97f25e6a9a2e3c89" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.370643 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c9eadefd0babc119f776d48c9d2c9e4bced58db685af9a8e97f25e6a9a2e3c89"} err="failed to get container status \"c9eadefd0babc119f776d48c9d2c9e4bced58db685af9a8e97f25e6a9a2e3c89\": rpc error: code = NotFound desc = could not find container \"c9eadefd0babc119f776d48c9d2c9e4bced58db685af9a8e97f25e6a9a2e3c89\": container with ID starting with c9eadefd0babc119f776d48c9d2c9e4bced58db685af9a8e97f25e6a9a2e3c89 not found: ID does not exist" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.370675 4948 scope.go:117] "RemoveContainer" containerID="84af4cdd213d90636a889a8a8a3aaca3dee1e2677368efa0ffc768940c2cc994" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.371058 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"84af4cdd213d90636a889a8a8a3aaca3dee1e2677368efa0ffc768940c2cc994"} err="failed to get container status \"84af4cdd213d90636a889a8a8a3aaca3dee1e2677368efa0ffc768940c2cc994\": rpc error: code = NotFound desc = could not find container \"84af4cdd213d90636a889a8a8a3aaca3dee1e2677368efa0ffc768940c2cc994\": container with ID starting with 84af4cdd213d90636a889a8a8a3aaca3dee1e2677368efa0ffc768940c2cc994 not found: ID does not exist" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.371099 4948 scope.go:117] "RemoveContainer" containerID="d4036d7131db00afd8eab8e093626b688297dc776b5af80c3c38b51da7124085" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.371600 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d4036d7131db00afd8eab8e093626b688297dc776b5af80c3c38b51da7124085"} err="failed to get container status \"d4036d7131db00afd8eab8e093626b688297dc776b5af80c3c38b51da7124085\": rpc error: code = NotFound desc = could not find container \"d4036d7131db00afd8eab8e093626b688297dc776b5af80c3c38b51da7124085\": container with ID starting with d4036d7131db00afd8eab8e093626b688297dc776b5af80c3c38b51da7124085 not found: ID does not exist" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.371630 4948 scope.go:117] "RemoveContainer" containerID="ec9d5e77753fda70350353daa0967257dd4929dba4440a95495ab4c2a0659aef" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.371967 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"ec9d5e77753fda70350353daa0967257dd4929dba4440a95495ab4c2a0659aef"} err="failed to get container status \"ec9d5e77753fda70350353daa0967257dd4929dba4440a95495ab4c2a0659aef\": rpc error: code = NotFound desc = could not find container \"ec9d5e77753fda70350353daa0967257dd4929dba4440a95495ab4c2a0659aef\": container with ID starting with ec9d5e77753fda70350353daa0967257dd4929dba4440a95495ab4c2a0659aef not found: ID does not exist" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.371987 4948 scope.go:117] "RemoveContainer" containerID="8cad943edb9a8d445b33a6e70d85b475c1ccba460a6056bccae94fe60fd6b73f" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.372462 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"8cad943edb9a8d445b33a6e70d85b475c1ccba460a6056bccae94fe60fd6b73f"} err="failed to get container status \"8cad943edb9a8d445b33a6e70d85b475c1ccba460a6056bccae94fe60fd6b73f\": rpc error: code = NotFound desc = could not find container \"8cad943edb9a8d445b33a6e70d85b475c1ccba460a6056bccae94fe60fd6b73f\": container with ID starting with 8cad943edb9a8d445b33a6e70d85b475c1ccba460a6056bccae94fe60fd6b73f not found: ID does not exist" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.372493 4948 scope.go:117] "RemoveContainer" containerID="c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.372762 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a"} err="failed to get container status \"c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\": rpc error: code = NotFound desc = could not find container \"c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a\": container with ID starting with c2cc477b22c9eed57350cea3247788ca3feda15b6a459dba80c69625d9815b7a not found: ID does not exist" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.978676 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-bwpqp_3b7580c2-5c23-4c67-807a-ea97a3df9398/kube-multus/2.log" Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.983644 4948 generic.go:334] "Generic (PLEG): container finished" podID="cb755fc0-e934-4bef-93e9-427cbb2ee692" containerID="bd9fad5911c671051c8aafbdcadb06f4fdf87b58555cd52925993775850ec69a" exitCode=0 Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.983737 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" event={"ID":"cb755fc0-e934-4bef-93e9-427cbb2ee692","Type":"ContainerDied","Data":"bd9fad5911c671051c8aafbdcadb06f4fdf87b58555cd52925993775850ec69a"} Mar 12 00:18:39 crc kubenswrapper[4948]: I0312 00:18:39.983827 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" event={"ID":"cb755fc0-e934-4bef-93e9-427cbb2ee692","Type":"ContainerStarted","Data":"cda20bfdbd21a11d7d53f5433e4b203039e1b3043a47420bb57af7aa2a30f9a4"} Mar 12 00:18:41 crc kubenswrapper[4948]: I0312 00:18:41.000852 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" event={"ID":"cb755fc0-e934-4bef-93e9-427cbb2ee692","Type":"ContainerStarted","Data":"9b0c49df8d394c05cbd11e95ff439234cd638c28801c5f3a272a1b90b7022bb2"} Mar 12 00:18:41 crc kubenswrapper[4948]: I0312 00:18:41.001155 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" event={"ID":"cb755fc0-e934-4bef-93e9-427cbb2ee692","Type":"ContainerStarted","Data":"0c274a6b78e2c28636e6526306274d45e2b7c19de69a35c7cffb66f4a5c56069"} Mar 12 00:18:41 crc kubenswrapper[4948]: I0312 00:18:41.001168 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" event={"ID":"cb755fc0-e934-4bef-93e9-427cbb2ee692","Type":"ContainerStarted","Data":"2fa2e879f1b98962e6a2a4c77da6fc304e8dbed02dc7bed5a4a9dd4385df6168"} Mar 12 00:18:41 crc kubenswrapper[4948]: I0312 00:18:41.001179 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" event={"ID":"cb755fc0-e934-4bef-93e9-427cbb2ee692","Type":"ContainerStarted","Data":"d2fa9a82c29f02f3dbabe9dd33a37ecb467787db8d6c2df05313b75803ca4c08"} Mar 12 00:18:41 crc kubenswrapper[4948]: I0312 00:18:41.001190 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" event={"ID":"cb755fc0-e934-4bef-93e9-427cbb2ee692","Type":"ContainerStarted","Data":"c634b5d2f3ee43812c751996489249f862b2fd3e2d003ce0fad3d2e7bf9cb8a6"} Mar 12 00:18:41 crc kubenswrapper[4948]: I0312 00:18:41.001202 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" event={"ID":"cb755fc0-e934-4bef-93e9-427cbb2ee692","Type":"ContainerStarted","Data":"abb4da4b2e6191f304452af92bac51e471a3e342031eaec10dee000749eaf1bd"} Mar 12 00:18:44 crc kubenswrapper[4948]: I0312 00:18:44.037044 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" event={"ID":"cb755fc0-e934-4bef-93e9-427cbb2ee692","Type":"ContainerStarted","Data":"42a3adf688657b986489f0caaf9b2f8206af9998151d374b9b91e3dfe6c20f2e"} Mar 12 00:18:46 crc kubenswrapper[4948]: I0312 00:18:46.055744 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" event={"ID":"cb755fc0-e934-4bef-93e9-427cbb2ee692","Type":"ContainerStarted","Data":"19ffaa3e21fd86118ce212c72578136d3629c07d24375d81f718bd7954745aed"} Mar 12 00:18:46 crc kubenswrapper[4948]: I0312 00:18:46.056094 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:18:46 crc kubenswrapper[4948]: I0312 00:18:46.056250 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:18:46 crc kubenswrapper[4948]: I0312 00:18:46.056298 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:18:46 crc kubenswrapper[4948]: I0312 00:18:46.082575 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:18:46 crc kubenswrapper[4948]: I0312 00:18:46.084883 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:18:46 crc kubenswrapper[4948]: I0312 00:18:46.102076 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" podStartSLOduration=8.102050988 podStartE2EDuration="8.102050988s" podCreationTimestamp="2026-03-12 00:18:38 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:18:46.096669363 +0000 UTC m=+725.552273101" watchObservedRunningTime="2026-03-12 00:18:46.102050988 +0000 UTC m=+725.557654746" Mar 12 00:18:51 crc kubenswrapper[4948]: I0312 00:18:51.317918 4948 scope.go:117] "RemoveContainer" containerID="b283d9461d9f6b7694a42e106d5d1c94fa66752ee64832f229a2e6088e3fca81" Mar 12 00:18:51 crc kubenswrapper[4948]: E0312 00:18:51.318952 4948 pod_workers.go:1301] "Error syncing pod, skipping" err="failed to \"StartContainer\" for \"kube-multus\" with CrashLoopBackOff: \"back-off 20s restarting failed container=kube-multus pod=multus-bwpqp_openshift-multus(3b7580c2-5c23-4c67-807a-ea97a3df9398)\"" pod="openshift-multus/multus-bwpqp" podUID="3b7580c2-5c23-4c67-807a-ea97a3df9398" Mar 12 00:19:02 crc kubenswrapper[4948]: I0312 00:19:02.312915 4948 scope.go:117] "RemoveContainer" containerID="b283d9461d9f6b7694a42e106d5d1c94fa66752ee64832f229a2e6088e3fca81" Mar 12 00:19:03 crc kubenswrapper[4948]: I0312 00:19:03.195454 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-multus_multus-bwpqp_3b7580c2-5c23-4c67-807a-ea97a3df9398/kube-multus/2.log" Mar 12 00:19:03 crc kubenswrapper[4948]: I0312 00:19:03.195837 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-multus/multus-bwpqp" event={"ID":"3b7580c2-5c23-4c67-807a-ea97a3df9398","Type":"ContainerStarted","Data":"1ba63f26dbd86b86fdc2834128c180adf4fd28a6ea1f369026434661f064da13"} Mar 12 00:19:08 crc kubenswrapper[4948]: I0312 00:19:08.948160 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-ovn-kubernetes/ovnkube-node-lqrtr" Mar 12 00:19:39 crc kubenswrapper[4948]: I0312 00:19:39.210679 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qqd6j"] Mar 12 00:19:39 crc kubenswrapper[4948]: I0312 00:19:39.211933 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-marketplace-qqd6j" podUID="67ac8dbf-c516-4b07-aa10-546e57acdf69" containerName="registry-server" containerID="cri-o://304db3562a72e39f59cfa721deef292badcb7adfdbf6b21f029f64c6001a1640" gracePeriod=30 Mar 12 00:19:39 crc kubenswrapper[4948]: I0312 00:19:39.444036 4948 generic.go:334] "Generic (PLEG): container finished" podID="67ac8dbf-c516-4b07-aa10-546e57acdf69" containerID="304db3562a72e39f59cfa721deef292badcb7adfdbf6b21f029f64c6001a1640" exitCode=0 Mar 12 00:19:39 crc kubenswrapper[4948]: I0312 00:19:39.444098 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qqd6j" event={"ID":"67ac8dbf-c516-4b07-aa10-546e57acdf69","Type":"ContainerDied","Data":"304db3562a72e39f59cfa721deef292badcb7adfdbf6b21f029f64c6001a1640"} Mar 12 00:19:39 crc kubenswrapper[4948]: I0312 00:19:39.682026 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qqd6j" Mar 12 00:19:39 crc kubenswrapper[4948]: I0312 00:19:39.790475 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67ac8dbf-c516-4b07-aa10-546e57acdf69-catalog-content\") pod \"67ac8dbf-c516-4b07-aa10-546e57acdf69\" (UID: \"67ac8dbf-c516-4b07-aa10-546e57acdf69\") " Mar 12 00:19:39 crc kubenswrapper[4948]: I0312 00:19:39.790566 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-s6x58\" (UniqueName: \"kubernetes.io/projected/67ac8dbf-c516-4b07-aa10-546e57acdf69-kube-api-access-s6x58\") pod \"67ac8dbf-c516-4b07-aa10-546e57acdf69\" (UID: \"67ac8dbf-c516-4b07-aa10-546e57acdf69\") " Mar 12 00:19:39 crc kubenswrapper[4948]: I0312 00:19:39.790633 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67ac8dbf-c516-4b07-aa10-546e57acdf69-utilities\") pod \"67ac8dbf-c516-4b07-aa10-546e57acdf69\" (UID: \"67ac8dbf-c516-4b07-aa10-546e57acdf69\") " Mar 12 00:19:39 crc kubenswrapper[4948]: I0312 00:19:39.791700 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/67ac8dbf-c516-4b07-aa10-546e57acdf69-utilities" (OuterVolumeSpecName: "utilities") pod "67ac8dbf-c516-4b07-aa10-546e57acdf69" (UID: "67ac8dbf-c516-4b07-aa10-546e57acdf69"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 00:19:39 crc kubenswrapper[4948]: I0312 00:19:39.797778 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/67ac8dbf-c516-4b07-aa10-546e57acdf69-kube-api-access-s6x58" (OuterVolumeSpecName: "kube-api-access-s6x58") pod "67ac8dbf-c516-4b07-aa10-546e57acdf69" (UID: "67ac8dbf-c516-4b07-aa10-546e57acdf69"). InnerVolumeSpecName "kube-api-access-s6x58". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:19:39 crc kubenswrapper[4948]: I0312 00:19:39.841047 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/67ac8dbf-c516-4b07-aa10-546e57acdf69-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "67ac8dbf-c516-4b07-aa10-546e57acdf69" (UID: "67ac8dbf-c516-4b07-aa10-546e57acdf69"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 00:19:39 crc kubenswrapper[4948]: I0312 00:19:39.891685 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/67ac8dbf-c516-4b07-aa10-546e57acdf69-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 00:19:39 crc kubenswrapper[4948]: I0312 00:19:39.891739 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-s6x58\" (UniqueName: \"kubernetes.io/projected/67ac8dbf-c516-4b07-aa10-546e57acdf69-kube-api-access-s6x58\") on node \"crc\" DevicePath \"\"" Mar 12 00:19:39 crc kubenswrapper[4948]: I0312 00:19:39.891766 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/67ac8dbf-c516-4b07-aa10-546e57acdf69-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 00:19:40 crc kubenswrapper[4948]: I0312 00:19:40.455755 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-marketplace-qqd6j" event={"ID":"67ac8dbf-c516-4b07-aa10-546e57acdf69","Type":"ContainerDied","Data":"8dc7620762f9f98691bcb63f014c023becf9e59ab861dba28deaef945fd60223"} Mar 12 00:19:40 crc kubenswrapper[4948]: I0312 00:19:40.455853 4948 scope.go:117] "RemoveContainer" containerID="304db3562a72e39f59cfa721deef292badcb7adfdbf6b21f029f64c6001a1640" Mar 12 00:19:40 crc kubenswrapper[4948]: I0312 00:19:40.456074 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-marketplace-qqd6j" Mar 12 00:19:40 crc kubenswrapper[4948]: I0312 00:19:40.486616 4948 scope.go:117] "RemoveContainer" containerID="a31ca9093f6c6f6d5a0728dee03e6c8a89881385ec908b95769d354322c96b6c" Mar 12 00:19:40 crc kubenswrapper[4948]: I0312 00:19:40.522155 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-marketplace-qqd6j"] Mar 12 00:19:40 crc kubenswrapper[4948]: I0312 00:19:40.522439 4948 scope.go:117] "RemoveContainer" containerID="6930ca7436e07739b40d3e73ef2dab180d025594762453c1b8ac6f0bacd45ef0" Mar 12 00:19:40 crc kubenswrapper[4948]: I0312 00:19:40.533091 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-marketplace-qqd6j"] Mar 12 00:19:41 crc kubenswrapper[4948]: I0312 00:19:41.334238 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="67ac8dbf-c516-4b07-aa10-546e57acdf69" path="/var/lib/kubelet/pods/67ac8dbf-c516-4b07-aa10-546e57acdf69/volumes" Mar 12 00:19:43 crc kubenswrapper[4948]: I0312 00:19:43.250413 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc676"] Mar 12 00:19:43 crc kubenswrapper[4948]: E0312 00:19:43.250824 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67ac8dbf-c516-4b07-aa10-546e57acdf69" containerName="registry-server" Mar 12 00:19:43 crc kubenswrapper[4948]: I0312 00:19:43.250848 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="67ac8dbf-c516-4b07-aa10-546e57acdf69" containerName="registry-server" Mar 12 00:19:43 crc kubenswrapper[4948]: E0312 00:19:43.250876 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67ac8dbf-c516-4b07-aa10-546e57acdf69" containerName="extract-content" Mar 12 00:19:43 crc kubenswrapper[4948]: I0312 00:19:43.250889 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="67ac8dbf-c516-4b07-aa10-546e57acdf69" containerName="extract-content" Mar 12 00:19:43 crc kubenswrapper[4948]: E0312 00:19:43.250914 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="67ac8dbf-c516-4b07-aa10-546e57acdf69" containerName="extract-utilities" Mar 12 00:19:43 crc kubenswrapper[4948]: I0312 00:19:43.250927 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="67ac8dbf-c516-4b07-aa10-546e57acdf69" containerName="extract-utilities" Mar 12 00:19:43 crc kubenswrapper[4948]: I0312 00:19:43.251166 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="67ac8dbf-c516-4b07-aa10-546e57acdf69" containerName="registry-server" Mar 12 00:19:43 crc kubenswrapper[4948]: I0312 00:19:43.252882 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc676" Mar 12 00:19:43 crc kubenswrapper[4948]: I0312 00:19:43.259208 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Mar 12 00:19:43 crc kubenswrapper[4948]: I0312 00:19:43.259728 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc676"] Mar 12 00:19:43 crc kubenswrapper[4948]: I0312 00:19:43.352258 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a2b1eac3-5989-4e06-b457-16991daaac81-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc676\" (UID: \"a2b1eac3-5989-4e06-b457-16991daaac81\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc676" Mar 12 00:19:43 crc kubenswrapper[4948]: I0312 00:19:43.352414 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8jb7r\" (UniqueName: \"kubernetes.io/projected/a2b1eac3-5989-4e06-b457-16991daaac81-kube-api-access-8jb7r\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc676\" (UID: \"a2b1eac3-5989-4e06-b457-16991daaac81\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc676" Mar 12 00:19:43 crc kubenswrapper[4948]: I0312 00:19:43.352501 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a2b1eac3-5989-4e06-b457-16991daaac81-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc676\" (UID: \"a2b1eac3-5989-4e06-b457-16991daaac81\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc676" Mar 12 00:19:43 crc kubenswrapper[4948]: I0312 00:19:43.453180 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a2b1eac3-5989-4e06-b457-16991daaac81-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc676\" (UID: \"a2b1eac3-5989-4e06-b457-16991daaac81\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc676" Mar 12 00:19:43 crc kubenswrapper[4948]: I0312 00:19:43.453280 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a2b1eac3-5989-4e06-b457-16991daaac81-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc676\" (UID: \"a2b1eac3-5989-4e06-b457-16991daaac81\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc676" Mar 12 00:19:43 crc kubenswrapper[4948]: I0312 00:19:43.453395 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8jb7r\" (UniqueName: \"kubernetes.io/projected/a2b1eac3-5989-4e06-b457-16991daaac81-kube-api-access-8jb7r\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc676\" (UID: \"a2b1eac3-5989-4e06-b457-16991daaac81\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc676" Mar 12 00:19:43 crc kubenswrapper[4948]: I0312 00:19:43.453820 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a2b1eac3-5989-4e06-b457-16991daaac81-util\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc676\" (UID: \"a2b1eac3-5989-4e06-b457-16991daaac81\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc676" Mar 12 00:19:43 crc kubenswrapper[4948]: I0312 00:19:43.454391 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a2b1eac3-5989-4e06-b457-16991daaac81-bundle\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc676\" (UID: \"a2b1eac3-5989-4e06-b457-16991daaac81\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc676" Mar 12 00:19:43 crc kubenswrapper[4948]: I0312 00:19:43.485763 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8jb7r\" (UniqueName: \"kubernetes.io/projected/a2b1eac3-5989-4e06-b457-16991daaac81-kube-api-access-8jb7r\") pod \"98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc676\" (UID: \"a2b1eac3-5989-4e06-b457-16991daaac81\") " pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc676" Mar 12 00:19:43 crc kubenswrapper[4948]: I0312 00:19:43.578995 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc676" Mar 12 00:19:43 crc kubenswrapper[4948]: I0312 00:19:43.858409 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc676"] Mar 12 00:19:43 crc kubenswrapper[4948]: W0312 00:19:43.873400 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-poda2b1eac3_5989_4e06_b457_16991daaac81.slice/crio-dcae189cf3f4ff2d80280a173f98b4d31c0c82e55827c1bf221a9683f35ed55a WatchSource:0}: Error finding container dcae189cf3f4ff2d80280a173f98b4d31c0c82e55827c1bf221a9683f35ed55a: Status 404 returned error can't find the container with id dcae189cf3f4ff2d80280a173f98b4d31c0c82e55827c1bf221a9683f35ed55a Mar 12 00:19:44 crc kubenswrapper[4948]: I0312 00:19:44.483157 4948 generic.go:334] "Generic (PLEG): container finished" podID="a2b1eac3-5989-4e06-b457-16991daaac81" containerID="d341f1a54024e6ddc81d60a56a4778a6445c81f3485142a41cf2bdf7aeaf9fc4" exitCode=0 Mar 12 00:19:44 crc kubenswrapper[4948]: I0312 00:19:44.483288 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc676" event={"ID":"a2b1eac3-5989-4e06-b457-16991daaac81","Type":"ContainerDied","Data":"d341f1a54024e6ddc81d60a56a4778a6445c81f3485142a41cf2bdf7aeaf9fc4"} Mar 12 00:19:44 crc kubenswrapper[4948]: I0312 00:19:44.483571 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc676" event={"ID":"a2b1eac3-5989-4e06-b457-16991daaac81","Type":"ContainerStarted","Data":"dcae189cf3f4ff2d80280a173f98b4d31c0c82e55827c1bf221a9683f35ed55a"} Mar 12 00:19:46 crc kubenswrapper[4948]: I0312 00:19:46.502556 4948 generic.go:334] "Generic (PLEG): container finished" podID="a2b1eac3-5989-4e06-b457-16991daaac81" containerID="2265873ad6f7da0813821d93f112da18fec078a8ea9a4c83d133bf20c0c56977" exitCode=0 Mar 12 00:19:46 crc kubenswrapper[4948]: I0312 00:19:46.502663 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc676" event={"ID":"a2b1eac3-5989-4e06-b457-16991daaac81","Type":"ContainerDied","Data":"2265873ad6f7da0813821d93f112da18fec078a8ea9a4c83d133bf20c0c56977"} Mar 12 00:19:47 crc kubenswrapper[4948]: I0312 00:19:47.515052 4948 generic.go:334] "Generic (PLEG): container finished" podID="a2b1eac3-5989-4e06-b457-16991daaac81" containerID="6ea9364b383863fc68ec43e2b5a8fcfc7a41446cd1f2f68221c4fb5ac3f6eb47" exitCode=0 Mar 12 00:19:47 crc kubenswrapper[4948]: I0312 00:19:47.515116 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc676" event={"ID":"a2b1eac3-5989-4e06-b457-16991daaac81","Type":"ContainerDied","Data":"6ea9364b383863fc68ec43e2b5a8fcfc7a41446cd1f2f68221c4fb5ac3f6eb47"} Mar 12 00:19:48 crc kubenswrapper[4948]: I0312 00:19:48.826768 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc676" Mar 12 00:19:49 crc kubenswrapper[4948]: I0312 00:19:49.026837 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-8jb7r\" (UniqueName: \"kubernetes.io/projected/a2b1eac3-5989-4e06-b457-16991daaac81-kube-api-access-8jb7r\") pod \"a2b1eac3-5989-4e06-b457-16991daaac81\" (UID: \"a2b1eac3-5989-4e06-b457-16991daaac81\") " Mar 12 00:19:49 crc kubenswrapper[4948]: I0312 00:19:49.026896 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a2b1eac3-5989-4e06-b457-16991daaac81-bundle\") pod \"a2b1eac3-5989-4e06-b457-16991daaac81\" (UID: \"a2b1eac3-5989-4e06-b457-16991daaac81\") " Mar 12 00:19:49 crc kubenswrapper[4948]: I0312 00:19:49.026999 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a2b1eac3-5989-4e06-b457-16991daaac81-util\") pod \"a2b1eac3-5989-4e06-b457-16991daaac81\" (UID: \"a2b1eac3-5989-4e06-b457-16991daaac81\") " Mar 12 00:19:49 crc kubenswrapper[4948]: I0312 00:19:49.031588 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a2b1eac3-5989-4e06-b457-16991daaac81-bundle" (OuterVolumeSpecName: "bundle") pod "a2b1eac3-5989-4e06-b457-16991daaac81" (UID: "a2b1eac3-5989-4e06-b457-16991daaac81"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 00:19:49 crc kubenswrapper[4948]: I0312 00:19:49.039673 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/a2b1eac3-5989-4e06-b457-16991daaac81-kube-api-access-8jb7r" (OuterVolumeSpecName: "kube-api-access-8jb7r") pod "a2b1eac3-5989-4e06-b457-16991daaac81" (UID: "a2b1eac3-5989-4e06-b457-16991daaac81"). InnerVolumeSpecName "kube-api-access-8jb7r". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:19:49 crc kubenswrapper[4948]: I0312 00:19:49.047657 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/a2b1eac3-5989-4e06-b457-16991daaac81-util" (OuterVolumeSpecName: "util") pod "a2b1eac3-5989-4e06-b457-16991daaac81" (UID: "a2b1eac3-5989-4e06-b457-16991daaac81"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 00:19:49 crc kubenswrapper[4948]: I0312 00:19:49.128939 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-8jb7r\" (UniqueName: \"kubernetes.io/projected/a2b1eac3-5989-4e06-b457-16991daaac81-kube-api-access-8jb7r\") on node \"crc\" DevicePath \"\"" Mar 12 00:19:49 crc kubenswrapper[4948]: I0312 00:19:49.128987 4948 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/a2b1eac3-5989-4e06-b457-16991daaac81-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 00:19:49 crc kubenswrapper[4948]: I0312 00:19:49.129002 4948 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/a2b1eac3-5989-4e06-b457-16991daaac81-util\") on node \"crc\" DevicePath \"\"" Mar 12 00:19:49 crc kubenswrapper[4948]: I0312 00:19:49.548056 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc676" event={"ID":"a2b1eac3-5989-4e06-b457-16991daaac81","Type":"ContainerDied","Data":"dcae189cf3f4ff2d80280a173f98b4d31c0c82e55827c1bf221a9683f35ed55a"} Mar 12 00:19:49 crc kubenswrapper[4948]: I0312 00:19:49.548120 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dcae189cf3f4ff2d80280a173f98b4d31c0c82e55827c1bf221a9683f35ed55a" Mar 12 00:19:49 crc kubenswrapper[4948]: I0312 00:19:49.548188 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc676" Mar 12 00:19:51 crc kubenswrapper[4948]: I0312 00:19:51.221125 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn6w4j"] Mar 12 00:19:51 crc kubenswrapper[4948]: E0312 00:19:51.221874 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2b1eac3-5989-4e06-b457-16991daaac81" containerName="pull" Mar 12 00:19:51 crc kubenswrapper[4948]: I0312 00:19:51.221891 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2b1eac3-5989-4e06-b457-16991daaac81" containerName="pull" Mar 12 00:19:51 crc kubenswrapper[4948]: E0312 00:19:51.221903 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2b1eac3-5989-4e06-b457-16991daaac81" containerName="util" Mar 12 00:19:51 crc kubenswrapper[4948]: I0312 00:19:51.221911 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2b1eac3-5989-4e06-b457-16991daaac81" containerName="util" Mar 12 00:19:51 crc kubenswrapper[4948]: E0312 00:19:51.221937 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="a2b1eac3-5989-4e06-b457-16991daaac81" containerName="extract" Mar 12 00:19:51 crc kubenswrapper[4948]: I0312 00:19:51.221947 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="a2b1eac3-5989-4e06-b457-16991daaac81" containerName="extract" Mar 12 00:19:51 crc kubenswrapper[4948]: I0312 00:19:51.222054 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="a2b1eac3-5989-4e06-b457-16991daaac81" containerName="extract" Mar 12 00:19:51 crc kubenswrapper[4948]: I0312 00:19:51.222969 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn6w4j" Mar 12 00:19:51 crc kubenswrapper[4948]: I0312 00:19:51.225976 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-marketplace"/"default-dockercfg-vmwhc" Mar 12 00:19:51 crc kubenswrapper[4948]: I0312 00:19:51.237850 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn6w4j"] Mar 12 00:19:51 crc kubenswrapper[4948]: I0312 00:19:51.366736 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d1dc348c-ee53-42f5-aebb-e482dff1ba79-util\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn6w4j\" (UID: \"d1dc348c-ee53-42f5-aebb-e482dff1ba79\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn6w4j" Mar 12 00:19:51 crc kubenswrapper[4948]: I0312 00:19:51.366823 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-f429f\" (UniqueName: \"kubernetes.io/projected/d1dc348c-ee53-42f5-aebb-e482dff1ba79-kube-api-access-f429f\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn6w4j\" (UID: \"d1dc348c-ee53-42f5-aebb-e482dff1ba79\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn6w4j" Mar 12 00:19:51 crc kubenswrapper[4948]: I0312 00:19:51.366909 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d1dc348c-ee53-42f5-aebb-e482dff1ba79-bundle\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn6w4j\" (UID: \"d1dc348c-ee53-42f5-aebb-e482dff1ba79\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn6w4j" Mar 12 00:19:51 crc kubenswrapper[4948]: I0312 00:19:51.468497 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d1dc348c-ee53-42f5-aebb-e482dff1ba79-bundle\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn6w4j\" (UID: \"d1dc348c-ee53-42f5-aebb-e482dff1ba79\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn6w4j" Mar 12 00:19:51 crc kubenswrapper[4948]: I0312 00:19:51.468601 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d1dc348c-ee53-42f5-aebb-e482dff1ba79-util\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn6w4j\" (UID: \"d1dc348c-ee53-42f5-aebb-e482dff1ba79\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn6w4j" Mar 12 00:19:51 crc kubenswrapper[4948]: I0312 00:19:51.468690 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-f429f\" (UniqueName: \"kubernetes.io/projected/d1dc348c-ee53-42f5-aebb-e482dff1ba79-kube-api-access-f429f\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn6w4j\" (UID: \"d1dc348c-ee53-42f5-aebb-e482dff1ba79\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn6w4j" Mar 12 00:19:51 crc kubenswrapper[4948]: I0312 00:19:51.469419 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d1dc348c-ee53-42f5-aebb-e482dff1ba79-bundle\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn6w4j\" (UID: \"d1dc348c-ee53-42f5-aebb-e482dff1ba79\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn6w4j" Mar 12 00:19:51 crc kubenswrapper[4948]: I0312 00:19:51.469984 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d1dc348c-ee53-42f5-aebb-e482dff1ba79-util\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn6w4j\" (UID: \"d1dc348c-ee53-42f5-aebb-e482dff1ba79\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn6w4j" Mar 12 00:19:51 crc kubenswrapper[4948]: I0312 00:19:51.504566 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-f429f\" (UniqueName: \"kubernetes.io/projected/d1dc348c-ee53-42f5-aebb-e482dff1ba79-kube-api-access-f429f\") pod \"6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn6w4j\" (UID: \"d1dc348c-ee53-42f5-aebb-e482dff1ba79\") " pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn6w4j" Mar 12 00:19:51 crc kubenswrapper[4948]: I0312 00:19:51.559143 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn6w4j" Mar 12 00:19:51 crc kubenswrapper[4948]: I0312 00:19:51.801095 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn6w4j"] Mar 12 00:19:52 crc kubenswrapper[4948]: I0312 00:19:52.002240 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39e9cm9g"] Mar 12 00:19:52 crc kubenswrapper[4948]: I0312 00:19:52.004806 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39e9cm9g" Mar 12 00:19:52 crc kubenswrapper[4948]: I0312 00:19:52.012552 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39e9cm9g"] Mar 12 00:19:52 crc kubenswrapper[4948]: I0312 00:19:52.076827 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/539fe10f-66ae-4110-a85e-0f42febbf096-bundle\") pod \"7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39e9cm9g\" (UID: \"539fe10f-66ae-4110-a85e-0f42febbf096\") " pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39e9cm9g" Mar 12 00:19:52 crc kubenswrapper[4948]: I0312 00:19:52.076895 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-t8tdp\" (UniqueName: \"kubernetes.io/projected/539fe10f-66ae-4110-a85e-0f42febbf096-kube-api-access-t8tdp\") pod \"7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39e9cm9g\" (UID: \"539fe10f-66ae-4110-a85e-0f42febbf096\") " pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39e9cm9g" Mar 12 00:19:52 crc kubenswrapper[4948]: I0312 00:19:52.076952 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/539fe10f-66ae-4110-a85e-0f42febbf096-util\") pod \"7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39e9cm9g\" (UID: \"539fe10f-66ae-4110-a85e-0f42febbf096\") " pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39e9cm9g" Mar 12 00:19:52 crc kubenswrapper[4948]: I0312 00:19:52.177995 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/539fe10f-66ae-4110-a85e-0f42febbf096-bundle\") pod \"7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39e9cm9g\" (UID: \"539fe10f-66ae-4110-a85e-0f42febbf096\") " pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39e9cm9g" Mar 12 00:19:52 crc kubenswrapper[4948]: I0312 00:19:52.178073 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-t8tdp\" (UniqueName: \"kubernetes.io/projected/539fe10f-66ae-4110-a85e-0f42febbf096-kube-api-access-t8tdp\") pod \"7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39e9cm9g\" (UID: \"539fe10f-66ae-4110-a85e-0f42febbf096\") " pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39e9cm9g" Mar 12 00:19:52 crc kubenswrapper[4948]: I0312 00:19:52.178154 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/539fe10f-66ae-4110-a85e-0f42febbf096-util\") pod \"7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39e9cm9g\" (UID: \"539fe10f-66ae-4110-a85e-0f42febbf096\") " pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39e9cm9g" Mar 12 00:19:52 crc kubenswrapper[4948]: I0312 00:19:52.179018 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/539fe10f-66ae-4110-a85e-0f42febbf096-util\") pod \"7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39e9cm9g\" (UID: \"539fe10f-66ae-4110-a85e-0f42febbf096\") " pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39e9cm9g" Mar 12 00:19:52 crc kubenswrapper[4948]: I0312 00:19:52.179062 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/539fe10f-66ae-4110-a85e-0f42febbf096-bundle\") pod \"7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39e9cm9g\" (UID: \"539fe10f-66ae-4110-a85e-0f42febbf096\") " pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39e9cm9g" Mar 12 00:19:52 crc kubenswrapper[4948]: I0312 00:19:52.205824 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-t8tdp\" (UniqueName: \"kubernetes.io/projected/539fe10f-66ae-4110-a85e-0f42febbf096-kube-api-access-t8tdp\") pod \"7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39e9cm9g\" (UID: \"539fe10f-66ae-4110-a85e-0f42febbf096\") " pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39e9cm9g" Mar 12 00:19:52 crc kubenswrapper[4948]: I0312 00:19:52.374926 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39e9cm9g" Mar 12 00:19:52 crc kubenswrapper[4948]: I0312 00:19:52.573504 4948 generic.go:334] "Generic (PLEG): container finished" podID="d1dc348c-ee53-42f5-aebb-e482dff1ba79" containerID="b664019bae6bfd60acf9e7dd80cdde690f0e0d00c1c40a0bfee45bfab933d445" exitCode=0 Mar 12 00:19:52 crc kubenswrapper[4948]: I0312 00:19:52.573575 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn6w4j" event={"ID":"d1dc348c-ee53-42f5-aebb-e482dff1ba79","Type":"ContainerDied","Data":"b664019bae6bfd60acf9e7dd80cdde690f0e0d00c1c40a0bfee45bfab933d445"} Mar 12 00:19:52 crc kubenswrapper[4948]: I0312 00:19:52.573983 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn6w4j" event={"ID":"d1dc348c-ee53-42f5-aebb-e482dff1ba79","Type":"ContainerStarted","Data":"fd9014f02aa1b74f665b3c947e449e6401ade2540d60ec9a159351b203555297"} Mar 12 00:19:52 crc kubenswrapper[4948]: I0312 00:19:52.648168 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39e9cm9g"] Mar 12 00:19:52 crc kubenswrapper[4948]: W0312 00:19:52.654783 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod539fe10f_66ae_4110_a85e_0f42febbf096.slice/crio-10b6239b328285f9ecd22426fe5e74f5297db574c7fccda2d020f52482b24d7e WatchSource:0}: Error finding container 10b6239b328285f9ecd22426fe5e74f5297db574c7fccda2d020f52482b24d7e: Status 404 returned error can't find the container with id 10b6239b328285f9ecd22426fe5e74f5297db574c7fccda2d020f52482b24d7e Mar 12 00:19:53 crc kubenswrapper[4948]: I0312 00:19:53.597119 4948 generic.go:334] "Generic (PLEG): container finished" podID="539fe10f-66ae-4110-a85e-0f42febbf096" containerID="2dbd0a533b03e229b0becdca0d2c1e1473a7ef337ea7127da5e3663447208d36" exitCode=0 Mar 12 00:19:53 crc kubenswrapper[4948]: I0312 00:19:53.597171 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39e9cm9g" event={"ID":"539fe10f-66ae-4110-a85e-0f42febbf096","Type":"ContainerDied","Data":"2dbd0a533b03e229b0becdca0d2c1e1473a7ef337ea7127da5e3663447208d36"} Mar 12 00:19:53 crc kubenswrapper[4948]: I0312 00:19:53.597202 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39e9cm9g" event={"ID":"539fe10f-66ae-4110-a85e-0f42febbf096","Type":"ContainerStarted","Data":"10b6239b328285f9ecd22426fe5e74f5297db574c7fccda2d020f52482b24d7e"} Mar 12 00:19:54 crc kubenswrapper[4948]: I0312 00:19:54.606949 4948 generic.go:334] "Generic (PLEG): container finished" podID="d1dc348c-ee53-42f5-aebb-e482dff1ba79" containerID="e3a0e064686df46944f1603b366d9304cafeee3b1cae5ee99b997ebd9dc2b947" exitCode=0 Mar 12 00:19:54 crc kubenswrapper[4948]: I0312 00:19:54.607019 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn6w4j" event={"ID":"d1dc348c-ee53-42f5-aebb-e482dff1ba79","Type":"ContainerDied","Data":"e3a0e064686df46944f1603b366d9304cafeee3b1cae5ee99b997ebd9dc2b947"} Mar 12 00:19:54 crc kubenswrapper[4948]: I0312 00:19:54.610384 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39e9cm9g" event={"ID":"539fe10f-66ae-4110-a85e-0f42febbf096","Type":"ContainerStarted","Data":"880ebf91c15714f4619305ecc3e323e4e8bf204e29e33a156073f5c1fae506bc"} Mar 12 00:19:55 crc kubenswrapper[4948]: I0312 00:19:55.617221 4948 generic.go:334] "Generic (PLEG): container finished" podID="539fe10f-66ae-4110-a85e-0f42febbf096" containerID="880ebf91c15714f4619305ecc3e323e4e8bf204e29e33a156073f5c1fae506bc" exitCode=0 Mar 12 00:19:55 crc kubenswrapper[4948]: I0312 00:19:55.617536 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39e9cm9g" event={"ID":"539fe10f-66ae-4110-a85e-0f42febbf096","Type":"ContainerDied","Data":"880ebf91c15714f4619305ecc3e323e4e8bf204e29e33a156073f5c1fae506bc"} Mar 12 00:19:55 crc kubenswrapper[4948]: I0312 00:19:55.621691 4948 generic.go:334] "Generic (PLEG): container finished" podID="d1dc348c-ee53-42f5-aebb-e482dff1ba79" containerID="940add7e6fbbe49f0595bd8795f08a0499e7ad8d183781630b4d50b3531360c8" exitCode=0 Mar 12 00:19:55 crc kubenswrapper[4948]: I0312 00:19:55.621731 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn6w4j" event={"ID":"d1dc348c-ee53-42f5-aebb-e482dff1ba79","Type":"ContainerDied","Data":"940add7e6fbbe49f0595bd8795f08a0499e7ad8d183781630b4d50b3531360c8"} Mar 12 00:19:55 crc kubenswrapper[4948]: I0312 00:19:55.658180 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5j9jsp"] Mar 12 00:19:55 crc kubenswrapper[4948]: I0312 00:19:55.659160 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5j9jsp" Mar 12 00:19:55 crc kubenswrapper[4948]: I0312 00:19:55.678392 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5j9jsp"] Mar 12 00:19:55 crc kubenswrapper[4948]: I0312 00:19:55.737856 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ncfd2\" (UniqueName: \"kubernetes.io/projected/3aa75b6b-35be-412c-8fa5-1e16b61a40fc-kube-api-access-ncfd2\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5j9jsp\" (UID: \"3aa75b6b-35be-412c-8fa5-1e16b61a40fc\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5j9jsp" Mar 12 00:19:55 crc kubenswrapper[4948]: I0312 00:19:55.738150 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3aa75b6b-35be-412c-8fa5-1e16b61a40fc-bundle\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5j9jsp\" (UID: \"3aa75b6b-35be-412c-8fa5-1e16b61a40fc\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5j9jsp" Mar 12 00:19:55 crc kubenswrapper[4948]: I0312 00:19:55.738211 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3aa75b6b-35be-412c-8fa5-1e16b61a40fc-util\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5j9jsp\" (UID: \"3aa75b6b-35be-412c-8fa5-1e16b61a40fc\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5j9jsp" Mar 12 00:19:55 crc kubenswrapper[4948]: I0312 00:19:55.839176 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ncfd2\" (UniqueName: \"kubernetes.io/projected/3aa75b6b-35be-412c-8fa5-1e16b61a40fc-kube-api-access-ncfd2\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5j9jsp\" (UID: \"3aa75b6b-35be-412c-8fa5-1e16b61a40fc\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5j9jsp" Mar 12 00:19:55 crc kubenswrapper[4948]: I0312 00:19:55.839235 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3aa75b6b-35be-412c-8fa5-1e16b61a40fc-bundle\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5j9jsp\" (UID: \"3aa75b6b-35be-412c-8fa5-1e16b61a40fc\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5j9jsp" Mar 12 00:19:55 crc kubenswrapper[4948]: I0312 00:19:55.839289 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3aa75b6b-35be-412c-8fa5-1e16b61a40fc-util\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5j9jsp\" (UID: \"3aa75b6b-35be-412c-8fa5-1e16b61a40fc\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5j9jsp" Mar 12 00:19:55 crc kubenswrapper[4948]: I0312 00:19:55.839675 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3aa75b6b-35be-412c-8fa5-1e16b61a40fc-util\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5j9jsp\" (UID: \"3aa75b6b-35be-412c-8fa5-1e16b61a40fc\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5j9jsp" Mar 12 00:19:55 crc kubenswrapper[4948]: I0312 00:19:55.839789 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3aa75b6b-35be-412c-8fa5-1e16b61a40fc-bundle\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5j9jsp\" (UID: \"3aa75b6b-35be-412c-8fa5-1e16b61a40fc\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5j9jsp" Mar 12 00:19:55 crc kubenswrapper[4948]: I0312 00:19:55.875209 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ncfd2\" (UniqueName: \"kubernetes.io/projected/3aa75b6b-35be-412c-8fa5-1e16b61a40fc-kube-api-access-ncfd2\") pod \"925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5j9jsp\" (UID: \"3aa75b6b-35be-412c-8fa5-1e16b61a40fc\") " pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5j9jsp" Mar 12 00:19:55 crc kubenswrapper[4948]: I0312 00:19:55.986212 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5j9jsp" Mar 12 00:19:56 crc kubenswrapper[4948]: I0312 00:19:56.367091 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5j9jsp"] Mar 12 00:19:56 crc kubenswrapper[4948]: W0312 00:19:56.370411 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod3aa75b6b_35be_412c_8fa5_1e16b61a40fc.slice/crio-8b9bc6ddbdbdb820665ed558943e3f305b8ea05474037a63d3669ad755a9b172 WatchSource:0}: Error finding container 8b9bc6ddbdbdb820665ed558943e3f305b8ea05474037a63d3669ad755a9b172: Status 404 returned error can't find the container with id 8b9bc6ddbdbdb820665ed558943e3f305b8ea05474037a63d3669ad755a9b172 Mar 12 00:19:56 crc kubenswrapper[4948]: I0312 00:19:56.629135 4948 generic.go:334] "Generic (PLEG): container finished" podID="539fe10f-66ae-4110-a85e-0f42febbf096" containerID="6ef178668ec397173de26f25107d9ccfc947fc6ec155c1129c8a64408d9d3117" exitCode=0 Mar 12 00:19:56 crc kubenswrapper[4948]: I0312 00:19:56.629577 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39e9cm9g" event={"ID":"539fe10f-66ae-4110-a85e-0f42febbf096","Type":"ContainerDied","Data":"6ef178668ec397173de26f25107d9ccfc947fc6ec155c1129c8a64408d9d3117"} Mar 12 00:19:56 crc kubenswrapper[4948]: I0312 00:19:56.631243 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5j9jsp" event={"ID":"3aa75b6b-35be-412c-8fa5-1e16b61a40fc","Type":"ContainerStarted","Data":"177d31c362ffec1e4494ebe00d4665d955b2b2fa95d10b5420d0faa7bc15036d"} Mar 12 00:19:56 crc kubenswrapper[4948]: I0312 00:19:56.631320 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5j9jsp" event={"ID":"3aa75b6b-35be-412c-8fa5-1e16b61a40fc","Type":"ContainerStarted","Data":"8b9bc6ddbdbdb820665ed558943e3f305b8ea05474037a63d3669ad755a9b172"} Mar 12 00:19:56 crc kubenswrapper[4948]: I0312 00:19:56.864011 4948 scope.go:117] "RemoveContainer" containerID="24742ed07ef4b5fa12b73123be5bed0fe00189424e312e3a45057377d14859df" Mar 12 00:19:56 crc kubenswrapper[4948]: I0312 00:19:56.974360 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn6w4j" Mar 12 00:19:57 crc kubenswrapper[4948]: I0312 00:19:57.053418 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-f429f\" (UniqueName: \"kubernetes.io/projected/d1dc348c-ee53-42f5-aebb-e482dff1ba79-kube-api-access-f429f\") pod \"d1dc348c-ee53-42f5-aebb-e482dff1ba79\" (UID: \"d1dc348c-ee53-42f5-aebb-e482dff1ba79\") " Mar 12 00:19:57 crc kubenswrapper[4948]: I0312 00:19:57.053580 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d1dc348c-ee53-42f5-aebb-e482dff1ba79-bundle\") pod \"d1dc348c-ee53-42f5-aebb-e482dff1ba79\" (UID: \"d1dc348c-ee53-42f5-aebb-e482dff1ba79\") " Mar 12 00:19:57 crc kubenswrapper[4948]: I0312 00:19:57.053674 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d1dc348c-ee53-42f5-aebb-e482dff1ba79-util\") pod \"d1dc348c-ee53-42f5-aebb-e482dff1ba79\" (UID: \"d1dc348c-ee53-42f5-aebb-e482dff1ba79\") " Mar 12 00:19:57 crc kubenswrapper[4948]: I0312 00:19:57.054014 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1dc348c-ee53-42f5-aebb-e482dff1ba79-bundle" (OuterVolumeSpecName: "bundle") pod "d1dc348c-ee53-42f5-aebb-e482dff1ba79" (UID: "d1dc348c-ee53-42f5-aebb-e482dff1ba79"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 00:19:57 crc kubenswrapper[4948]: I0312 00:19:57.054139 4948 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/d1dc348c-ee53-42f5-aebb-e482dff1ba79-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 00:19:57 crc kubenswrapper[4948]: I0312 00:19:57.071331 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/d1dc348c-ee53-42f5-aebb-e482dff1ba79-kube-api-access-f429f" (OuterVolumeSpecName: "kube-api-access-f429f") pod "d1dc348c-ee53-42f5-aebb-e482dff1ba79" (UID: "d1dc348c-ee53-42f5-aebb-e482dff1ba79"). InnerVolumeSpecName "kube-api-access-f429f". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:19:57 crc kubenswrapper[4948]: I0312 00:19:57.155862 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-f429f\" (UniqueName: \"kubernetes.io/projected/d1dc348c-ee53-42f5-aebb-e482dff1ba79-kube-api-access-f429f\") on node \"crc\" DevicePath \"\"" Mar 12 00:19:57 crc kubenswrapper[4948]: I0312 00:19:57.324342 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/d1dc348c-ee53-42f5-aebb-e482dff1ba79-util" (OuterVolumeSpecName: "util") pod "d1dc348c-ee53-42f5-aebb-e482dff1ba79" (UID: "d1dc348c-ee53-42f5-aebb-e482dff1ba79"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 00:19:57 crc kubenswrapper[4948]: I0312 00:19:57.358731 4948 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/d1dc348c-ee53-42f5-aebb-e482dff1ba79-util\") on node \"crc\" DevicePath \"\"" Mar 12 00:19:57 crc kubenswrapper[4948]: I0312 00:19:57.637447 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn6w4j" event={"ID":"d1dc348c-ee53-42f5-aebb-e482dff1ba79","Type":"ContainerDied","Data":"fd9014f02aa1b74f665b3c947e449e6401ade2540d60ec9a159351b203555297"} Mar 12 00:19:57 crc kubenswrapper[4948]: I0312 00:19:57.637480 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="fd9014f02aa1b74f665b3c947e449e6401ade2540d60ec9a159351b203555297" Mar 12 00:19:57 crc kubenswrapper[4948]: I0312 00:19:57.637492 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn6w4j" Mar 12 00:19:57 crc kubenswrapper[4948]: I0312 00:19:57.638597 4948 generic.go:334] "Generic (PLEG): container finished" podID="3aa75b6b-35be-412c-8fa5-1e16b61a40fc" containerID="177d31c362ffec1e4494ebe00d4665d955b2b2fa95d10b5420d0faa7bc15036d" exitCode=0 Mar 12 00:19:57 crc kubenswrapper[4948]: I0312 00:19:57.639259 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5j9jsp" event={"ID":"3aa75b6b-35be-412c-8fa5-1e16b61a40fc","Type":"ContainerDied","Data":"177d31c362ffec1e4494ebe00d4665d955b2b2fa95d10b5420d0faa7bc15036d"} Mar 12 00:19:58 crc kubenswrapper[4948]: I0312 00:19:58.020650 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39e9cm9g" Mar 12 00:19:58 crc kubenswrapper[4948]: I0312 00:19:58.066097 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/539fe10f-66ae-4110-a85e-0f42febbf096-util\") pod \"539fe10f-66ae-4110-a85e-0f42febbf096\" (UID: \"539fe10f-66ae-4110-a85e-0f42febbf096\") " Mar 12 00:19:58 crc kubenswrapper[4948]: I0312 00:19:58.066143 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/539fe10f-66ae-4110-a85e-0f42febbf096-bundle\") pod \"539fe10f-66ae-4110-a85e-0f42febbf096\" (UID: \"539fe10f-66ae-4110-a85e-0f42febbf096\") " Mar 12 00:19:58 crc kubenswrapper[4948]: I0312 00:19:58.066234 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-t8tdp\" (UniqueName: \"kubernetes.io/projected/539fe10f-66ae-4110-a85e-0f42febbf096-kube-api-access-t8tdp\") pod \"539fe10f-66ae-4110-a85e-0f42febbf096\" (UID: \"539fe10f-66ae-4110-a85e-0f42febbf096\") " Mar 12 00:19:58 crc kubenswrapper[4948]: I0312 00:19:58.066991 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/539fe10f-66ae-4110-a85e-0f42febbf096-bundle" (OuterVolumeSpecName: "bundle") pod "539fe10f-66ae-4110-a85e-0f42febbf096" (UID: "539fe10f-66ae-4110-a85e-0f42febbf096"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 00:19:58 crc kubenswrapper[4948]: I0312 00:19:58.070398 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/539fe10f-66ae-4110-a85e-0f42febbf096-kube-api-access-t8tdp" (OuterVolumeSpecName: "kube-api-access-t8tdp") pod "539fe10f-66ae-4110-a85e-0f42febbf096" (UID: "539fe10f-66ae-4110-a85e-0f42febbf096"). InnerVolumeSpecName "kube-api-access-t8tdp". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:19:58 crc kubenswrapper[4948]: I0312 00:19:58.096542 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/539fe10f-66ae-4110-a85e-0f42febbf096-util" (OuterVolumeSpecName: "util") pod "539fe10f-66ae-4110-a85e-0f42febbf096" (UID: "539fe10f-66ae-4110-a85e-0f42febbf096"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 00:19:58 crc kubenswrapper[4948]: I0312 00:19:58.168017 4948 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/539fe10f-66ae-4110-a85e-0f42febbf096-util\") on node \"crc\" DevicePath \"\"" Mar 12 00:19:58 crc kubenswrapper[4948]: I0312 00:19:58.168229 4948 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/539fe10f-66ae-4110-a85e-0f42febbf096-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 00:19:58 crc kubenswrapper[4948]: I0312 00:19:58.168239 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-t8tdp\" (UniqueName: \"kubernetes.io/projected/539fe10f-66ae-4110-a85e-0f42febbf096-kube-api-access-t8tdp\") on node \"crc\" DevicePath \"\"" Mar 12 00:19:58 crc kubenswrapper[4948]: I0312 00:19:58.647743 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39e9cm9g" event={"ID":"539fe10f-66ae-4110-a85e-0f42febbf096","Type":"ContainerDied","Data":"10b6239b328285f9ecd22426fe5e74f5297db574c7fccda2d020f52482b24d7e"} Mar 12 00:19:58 crc kubenswrapper[4948]: I0312 00:19:58.647779 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="10b6239b328285f9ecd22426fe5e74f5297db574c7fccda2d020f52482b24d7e" Mar 12 00:19:58 crc kubenswrapper[4948]: I0312 00:19:58.647833 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39e9cm9g" Mar 12 00:20:00 crc kubenswrapper[4948]: I0312 00:20:00.236347 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29554580-ghvkh"] Mar 12 00:20:00 crc kubenswrapper[4948]: E0312 00:20:00.237436 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1dc348c-ee53-42f5-aebb-e482dff1ba79" containerName="pull" Mar 12 00:20:00 crc kubenswrapper[4948]: I0312 00:20:00.237537 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1dc348c-ee53-42f5-aebb-e482dff1ba79" containerName="pull" Mar 12 00:20:00 crc kubenswrapper[4948]: E0312 00:20:00.237641 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1dc348c-ee53-42f5-aebb-e482dff1ba79" containerName="extract" Mar 12 00:20:00 crc kubenswrapper[4948]: I0312 00:20:00.237725 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1dc348c-ee53-42f5-aebb-e482dff1ba79" containerName="extract" Mar 12 00:20:00 crc kubenswrapper[4948]: E0312 00:20:00.237797 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="539fe10f-66ae-4110-a85e-0f42febbf096" containerName="pull" Mar 12 00:20:00 crc kubenswrapper[4948]: I0312 00:20:00.237864 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="539fe10f-66ae-4110-a85e-0f42febbf096" containerName="pull" Mar 12 00:20:00 crc kubenswrapper[4948]: E0312 00:20:00.237934 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="d1dc348c-ee53-42f5-aebb-e482dff1ba79" containerName="util" Mar 12 00:20:00 crc kubenswrapper[4948]: I0312 00:20:00.238002 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="d1dc348c-ee53-42f5-aebb-e482dff1ba79" containerName="util" Mar 12 00:20:00 crc kubenswrapper[4948]: E0312 00:20:00.238074 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="539fe10f-66ae-4110-a85e-0f42febbf096" containerName="util" Mar 12 00:20:00 crc kubenswrapper[4948]: I0312 00:20:00.238137 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="539fe10f-66ae-4110-a85e-0f42febbf096" containerName="util" Mar 12 00:20:00 crc kubenswrapper[4948]: E0312 00:20:00.238204 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="539fe10f-66ae-4110-a85e-0f42febbf096" containerName="extract" Mar 12 00:20:00 crc kubenswrapper[4948]: I0312 00:20:00.238273 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="539fe10f-66ae-4110-a85e-0f42febbf096" containerName="extract" Mar 12 00:20:00 crc kubenswrapper[4948]: I0312 00:20:00.238501 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="d1dc348c-ee53-42f5-aebb-e482dff1ba79" containerName="extract" Mar 12 00:20:00 crc kubenswrapper[4948]: I0312 00:20:00.238590 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="539fe10f-66ae-4110-a85e-0f42febbf096" containerName="extract" Mar 12 00:20:00 crc kubenswrapper[4948]: I0312 00:20:00.239102 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29554580-ghvkh" Mar 12 00:20:00 crc kubenswrapper[4948]: I0312 00:20:00.242106 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 00:20:00 crc kubenswrapper[4948]: I0312 00:20:00.242141 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 00:20:00 crc kubenswrapper[4948]: I0312 00:20:00.242409 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-44g5m" Mar 12 00:20:00 crc kubenswrapper[4948]: I0312 00:20:00.250369 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29554580-ghvkh"] Mar 12 00:20:00 crc kubenswrapper[4948]: I0312 00:20:00.293176 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4wb5d\" (UniqueName: \"kubernetes.io/projected/4696b53e-388b-457f-9dc9-93811078ccc4-kube-api-access-4wb5d\") pod \"auto-csr-approver-29554580-ghvkh\" (UID: \"4696b53e-388b-457f-9dc9-93811078ccc4\") " pod="openshift-infra/auto-csr-approver-29554580-ghvkh" Mar 12 00:20:00 crc kubenswrapper[4948]: I0312 00:20:00.396038 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4wb5d\" (UniqueName: \"kubernetes.io/projected/4696b53e-388b-457f-9dc9-93811078ccc4-kube-api-access-4wb5d\") pod \"auto-csr-approver-29554580-ghvkh\" (UID: \"4696b53e-388b-457f-9dc9-93811078ccc4\") " pod="openshift-infra/auto-csr-approver-29554580-ghvkh" Mar 12 00:20:00 crc kubenswrapper[4948]: I0312 00:20:00.423552 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4wb5d\" (UniqueName: \"kubernetes.io/projected/4696b53e-388b-457f-9dc9-93811078ccc4-kube-api-access-4wb5d\") pod \"auto-csr-approver-29554580-ghvkh\" (UID: \"4696b53e-388b-457f-9dc9-93811078ccc4\") " pod="openshift-infra/auto-csr-approver-29554580-ghvkh" Mar 12 00:20:00 crc kubenswrapper[4948]: I0312 00:20:00.557744 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29554580-ghvkh" Mar 12 00:20:01 crc kubenswrapper[4948]: I0312 00:20:01.160092 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-cvw7f"] Mar 12 00:20:01 crc kubenswrapper[4948]: I0312 00:20:01.160695 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-cvw7f" Mar 12 00:20:01 crc kubenswrapper[4948]: I0312 00:20:01.162388 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-dockercfg-24qwg" Mar 12 00:20:01 crc kubenswrapper[4948]: I0312 00:20:01.162929 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"kube-root-ca.crt" Mar 12 00:20:01 crc kubenswrapper[4948]: I0312 00:20:01.163468 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-operators"/"openshift-service-ca.crt" Mar 12 00:20:01 crc kubenswrapper[4948]: I0312 00:20:01.171210 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-cvw7f"] Mar 12 00:20:01 crc kubenswrapper[4948]: I0312 00:20:01.209999 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xq7b4\" (UniqueName: \"kubernetes.io/projected/b5f03616-ebd8-4698-91f7-ca2419301e08-kube-api-access-xq7b4\") pod \"obo-prometheus-operator-68bc856cb9-cvw7f\" (UID: \"b5f03616-ebd8-4698-91f7-ca2419301e08\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-cvw7f" Mar 12 00:20:01 crc kubenswrapper[4948]: I0312 00:20:01.279058 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-85fd558b5-d55ck"] Mar 12 00:20:01 crc kubenswrapper[4948]: I0312 00:20:01.279705 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-85fd558b5-d55ck" Mar 12 00:20:01 crc kubenswrapper[4948]: I0312 00:20:01.281591 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-service-cert" Mar 12 00:20:01 crc kubenswrapper[4948]: I0312 00:20:01.281890 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"obo-prometheus-operator-admission-webhook-dockercfg-xmrw8" Mar 12 00:20:01 crc kubenswrapper[4948]: I0312 00:20:01.296102 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-85fd558b5-jvl2f"] Mar 12 00:20:01 crc kubenswrapper[4948]: I0312 00:20:01.296758 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-85fd558b5-jvl2f" Mar 12 00:20:01 crc kubenswrapper[4948]: I0312 00:20:01.303985 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-85fd558b5-d55ck"] Mar 12 00:20:01 crc kubenswrapper[4948]: I0312 00:20:01.311997 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/06125956-526b-468b-b609-9bb13c834499-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-85fd558b5-d55ck\" (UID: \"06125956-526b-468b-b609-9bb13c834499\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-85fd558b5-d55ck" Mar 12 00:20:01 crc kubenswrapper[4948]: I0312 00:20:01.312043 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xq7b4\" (UniqueName: \"kubernetes.io/projected/b5f03616-ebd8-4698-91f7-ca2419301e08-kube-api-access-xq7b4\") pod \"obo-prometheus-operator-68bc856cb9-cvw7f\" (UID: \"b5f03616-ebd8-4698-91f7-ca2419301e08\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-cvw7f" Mar 12 00:20:01 crc kubenswrapper[4948]: I0312 00:20:01.312090 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/06125956-526b-468b-b609-9bb13c834499-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-85fd558b5-d55ck\" (UID: \"06125956-526b-468b-b609-9bb13c834499\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-85fd558b5-d55ck" Mar 12 00:20:01 crc kubenswrapper[4948]: I0312 00:20:01.324627 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-85fd558b5-jvl2f"] Mar 12 00:20:01 crc kubenswrapper[4948]: I0312 00:20:01.329770 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xq7b4\" (UniqueName: \"kubernetes.io/projected/b5f03616-ebd8-4698-91f7-ca2419301e08-kube-api-access-xq7b4\") pod \"obo-prometheus-operator-68bc856cb9-cvw7f\" (UID: \"b5f03616-ebd8-4698-91f7-ca2419301e08\") " pod="openshift-operators/obo-prometheus-operator-68bc856cb9-cvw7f" Mar 12 00:20:01 crc kubenswrapper[4948]: I0312 00:20:01.412966 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/06125956-526b-468b-b609-9bb13c834499-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-85fd558b5-d55ck\" (UID: \"06125956-526b-468b-b609-9bb13c834499\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-85fd558b5-d55ck" Mar 12 00:20:01 crc kubenswrapper[4948]: I0312 00:20:01.413034 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/406df0e7-9930-440b-9023-5369cdac443b-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-85fd558b5-jvl2f\" (UID: \"406df0e7-9930-440b-9023-5369cdac443b\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-85fd558b5-jvl2f" Mar 12 00:20:01 crc kubenswrapper[4948]: I0312 00:20:01.413103 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/06125956-526b-468b-b609-9bb13c834499-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-85fd558b5-d55ck\" (UID: \"06125956-526b-468b-b609-9bb13c834499\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-85fd558b5-d55ck" Mar 12 00:20:01 crc kubenswrapper[4948]: I0312 00:20:01.413150 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/406df0e7-9930-440b-9023-5369cdac443b-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-85fd558b5-jvl2f\" (UID: \"406df0e7-9930-440b-9023-5369cdac443b\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-85fd558b5-jvl2f" Mar 12 00:20:01 crc kubenswrapper[4948]: I0312 00:20:01.416397 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/06125956-526b-468b-b609-9bb13c834499-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-85fd558b5-d55ck\" (UID: \"06125956-526b-468b-b609-9bb13c834499\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-85fd558b5-d55ck" Mar 12 00:20:01 crc kubenswrapper[4948]: I0312 00:20:01.416784 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/06125956-526b-468b-b609-9bb13c834499-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-85fd558b5-d55ck\" (UID: \"06125956-526b-468b-b609-9bb13c834499\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-85fd558b5-d55ck" Mar 12 00:20:01 crc kubenswrapper[4948]: I0312 00:20:01.480911 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-cvw7f" Mar 12 00:20:01 crc kubenswrapper[4948]: I0312 00:20:01.514472 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/406df0e7-9930-440b-9023-5369cdac443b-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-85fd558b5-jvl2f\" (UID: \"406df0e7-9930-440b-9023-5369cdac443b\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-85fd558b5-jvl2f" Mar 12 00:20:01 crc kubenswrapper[4948]: I0312 00:20:01.514567 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/406df0e7-9930-440b-9023-5369cdac443b-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-85fd558b5-jvl2f\" (UID: \"406df0e7-9930-440b-9023-5369cdac443b\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-85fd558b5-jvl2f" Mar 12 00:20:01 crc kubenswrapper[4948]: I0312 00:20:01.517497 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/406df0e7-9930-440b-9023-5369cdac443b-apiservice-cert\") pod \"obo-prometheus-operator-admission-webhook-85fd558b5-jvl2f\" (UID: \"406df0e7-9930-440b-9023-5369cdac443b\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-85fd558b5-jvl2f" Mar 12 00:20:01 crc kubenswrapper[4948]: I0312 00:20:01.521766 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/406df0e7-9930-440b-9023-5369cdac443b-webhook-cert\") pod \"obo-prometheus-operator-admission-webhook-85fd558b5-jvl2f\" (UID: \"406df0e7-9930-440b-9023-5369cdac443b\") " pod="openshift-operators/obo-prometheus-operator-admission-webhook-85fd558b5-jvl2f" Mar 12 00:20:01 crc kubenswrapper[4948]: I0312 00:20:01.546327 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-bkzp4"] Mar 12 00:20:01 crc kubenswrapper[4948]: I0312 00:20:01.546932 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-59bdc8b94-bkzp4" Mar 12 00:20:01 crc kubenswrapper[4948]: I0312 00:20:01.550035 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-tls" Mar 12 00:20:01 crc kubenswrapper[4948]: I0312 00:20:01.550354 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"observability-operator-sa-dockercfg-wz7wg" Mar 12 00:20:01 crc kubenswrapper[4948]: I0312 00:20:01.566221 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-bkzp4"] Mar 12 00:20:01 crc kubenswrapper[4948]: I0312 00:20:01.610649 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-85fd558b5-d55ck" Mar 12 00:20:01 crc kubenswrapper[4948]: I0312 00:20:01.611744 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/obo-prometheus-operator-admission-webhook-85fd558b5-jvl2f" Mar 12 00:20:01 crc kubenswrapper[4948]: I0312 00:20:01.615227 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf64ec59-b44a-407b-8d10-f1f9bd91de76-observability-operator-tls\") pod \"observability-operator-59bdc8b94-bkzp4\" (UID: \"bf64ec59-b44a-407b-8d10-f1f9bd91de76\") " pod="openshift-operators/observability-operator-59bdc8b94-bkzp4" Mar 12 00:20:01 crc kubenswrapper[4948]: I0312 00:20:01.615286 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-8h8rt\" (UniqueName: \"kubernetes.io/projected/bf64ec59-b44a-407b-8d10-f1f9bd91de76-kube-api-access-8h8rt\") pod \"observability-operator-59bdc8b94-bkzp4\" (UID: \"bf64ec59-b44a-407b-8d10-f1f9bd91de76\") " pod="openshift-operators/observability-operator-59bdc8b94-bkzp4" Mar 12 00:20:01 crc kubenswrapper[4948]: I0312 00:20:01.724212 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-swd7r"] Mar 12 00:20:01 crc kubenswrapper[4948]: I0312 00:20:01.724941 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5bf474d74f-swd7r" Mar 12 00:20:01 crc kubenswrapper[4948]: I0312 00:20:01.727211 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-operators"/"perses-operator-dockercfg-sbplx" Mar 12 00:20:01 crc kubenswrapper[4948]: I0312 00:20:01.730337 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29554580-ghvkh"] Mar 12 00:20:01 crc kubenswrapper[4948]: I0312 00:20:01.732774 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf64ec59-b44a-407b-8d10-f1f9bd91de76-observability-operator-tls\") pod \"observability-operator-59bdc8b94-bkzp4\" (UID: \"bf64ec59-b44a-407b-8d10-f1f9bd91de76\") " pod="openshift-operators/observability-operator-59bdc8b94-bkzp4" Mar 12 00:20:01 crc kubenswrapper[4948]: I0312 00:20:01.732847 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-8h8rt\" (UniqueName: \"kubernetes.io/projected/bf64ec59-b44a-407b-8d10-f1f9bd91de76-kube-api-access-8h8rt\") pod \"observability-operator-59bdc8b94-bkzp4\" (UID: \"bf64ec59-b44a-407b-8d10-f1f9bd91de76\") " pod="openshift-operators/observability-operator-59bdc8b94-bkzp4" Mar 12 00:20:01 crc kubenswrapper[4948]: I0312 00:20:01.738163 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-swd7r"] Mar 12 00:20:01 crc kubenswrapper[4948]: I0312 00:20:01.747467 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"observability-operator-tls\" (UniqueName: \"kubernetes.io/secret/bf64ec59-b44a-407b-8d10-f1f9bd91de76-observability-operator-tls\") pod \"observability-operator-59bdc8b94-bkzp4\" (UID: \"bf64ec59-b44a-407b-8d10-f1f9bd91de76\") " pod="openshift-operators/observability-operator-59bdc8b94-bkzp4" Mar 12 00:20:01 crc kubenswrapper[4948]: W0312 00:20:01.748545 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod4696b53e_388b_457f_9dc9_93811078ccc4.slice/crio-a7440a5c939651f60fea120b345322a008a40a836b1b833507aac8feb7decb6a WatchSource:0}: Error finding container a7440a5c939651f60fea120b345322a008a40a836b1b833507aac8feb7decb6a: Status 404 returned error can't find the container with id a7440a5c939651f60fea120b345322a008a40a836b1b833507aac8feb7decb6a Mar 12 00:20:01 crc kubenswrapper[4948]: I0312 00:20:01.752410 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-8h8rt\" (UniqueName: \"kubernetes.io/projected/bf64ec59-b44a-407b-8d10-f1f9bd91de76-kube-api-access-8h8rt\") pod \"observability-operator-59bdc8b94-bkzp4\" (UID: \"bf64ec59-b44a-407b-8d10-f1f9bd91de76\") " pod="openshift-operators/observability-operator-59bdc8b94-bkzp4" Mar 12 00:20:01 crc kubenswrapper[4948]: I0312 00:20:01.834828 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-ndr2k\" (UniqueName: \"kubernetes.io/projected/f7223c5e-0d07-4513-a05a-3d504647724c-kube-api-access-ndr2k\") pod \"perses-operator-5bf474d74f-swd7r\" (UID: \"f7223c5e-0d07-4513-a05a-3d504647724c\") " pod="openshift-operators/perses-operator-5bf474d74f-swd7r" Mar 12 00:20:01 crc kubenswrapper[4948]: I0312 00:20:01.837613 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/f7223c5e-0d07-4513-a05a-3d504647724c-openshift-service-ca\") pod \"perses-operator-5bf474d74f-swd7r\" (UID: \"f7223c5e-0d07-4513-a05a-3d504647724c\") " pod="openshift-operators/perses-operator-5bf474d74f-swd7r" Mar 12 00:20:01 crc kubenswrapper[4948]: I0312 00:20:01.920586 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/observability-operator-59bdc8b94-bkzp4" Mar 12 00:20:01 crc kubenswrapper[4948]: I0312 00:20:01.939134 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-ndr2k\" (UniqueName: \"kubernetes.io/projected/f7223c5e-0d07-4513-a05a-3d504647724c-kube-api-access-ndr2k\") pod \"perses-operator-5bf474d74f-swd7r\" (UID: \"f7223c5e-0d07-4513-a05a-3d504647724c\") " pod="openshift-operators/perses-operator-5bf474d74f-swd7r" Mar 12 00:20:01 crc kubenswrapper[4948]: I0312 00:20:01.939543 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/f7223c5e-0d07-4513-a05a-3d504647724c-openshift-service-ca\") pod \"perses-operator-5bf474d74f-swd7r\" (UID: \"f7223c5e-0d07-4513-a05a-3d504647724c\") " pod="openshift-operators/perses-operator-5bf474d74f-swd7r" Mar 12 00:20:01 crc kubenswrapper[4948]: I0312 00:20:01.940589 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"openshift-service-ca\" (UniqueName: \"kubernetes.io/configmap/f7223c5e-0d07-4513-a05a-3d504647724c-openshift-service-ca\") pod \"perses-operator-5bf474d74f-swd7r\" (UID: \"f7223c5e-0d07-4513-a05a-3d504647724c\") " pod="openshift-operators/perses-operator-5bf474d74f-swd7r" Mar 12 00:20:01 crc kubenswrapper[4948]: I0312 00:20:01.992600 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-ndr2k\" (UniqueName: \"kubernetes.io/projected/f7223c5e-0d07-4513-a05a-3d504647724c-kube-api-access-ndr2k\") pod \"perses-operator-5bf474d74f-swd7r\" (UID: \"f7223c5e-0d07-4513-a05a-3d504647724c\") " pod="openshift-operators/perses-operator-5bf474d74f-swd7r" Mar 12 00:20:01 crc kubenswrapper[4948]: I0312 00:20:01.998228 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-85fd558b5-d55ck"] Mar 12 00:20:02 crc kubenswrapper[4948]: I0312 00:20:02.053566 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-operators/perses-operator-5bf474d74f-swd7r" Mar 12 00:20:02 crc kubenswrapper[4948]: I0312 00:20:02.071330 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-68bc856cb9-cvw7f"] Mar 12 00:20:02 crc kubenswrapper[4948]: I0312 00:20:02.135582 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/observability-operator-59bdc8b94-bkzp4"] Mar 12 00:20:02 crc kubenswrapper[4948]: I0312 00:20:02.146568 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/obo-prometheus-operator-admission-webhook-85fd558b5-jvl2f"] Mar 12 00:20:02 crc kubenswrapper[4948]: W0312 00:20:02.157931 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-podbf64ec59_b44a_407b_8d10_f1f9bd91de76.slice/crio-5af6cb8851038058e53cb229ef7018840c49dc87c6681b26f0247ae679ab4272 WatchSource:0}: Error finding container 5af6cb8851038058e53cb229ef7018840c49dc87c6681b26f0247ae679ab4272: Status 404 returned error can't find the container with id 5af6cb8851038058e53cb229ef7018840c49dc87c6681b26f0247ae679ab4272 Mar 12 00:20:02 crc kubenswrapper[4948]: I0312 00:20:02.271008 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-operators/perses-operator-5bf474d74f-swd7r"] Mar 12 00:20:02 crc kubenswrapper[4948]: I0312 00:20:02.703403 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-59bdc8b94-bkzp4" event={"ID":"bf64ec59-b44a-407b-8d10-f1f9bd91de76","Type":"ContainerStarted","Data":"5af6cb8851038058e53cb229ef7018840c49dc87c6681b26f0247ae679ab4272"} Mar 12 00:20:02 crc kubenswrapper[4948]: I0312 00:20:02.706217 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-85fd558b5-d55ck" event={"ID":"06125956-526b-468b-b609-9bb13c834499","Type":"ContainerStarted","Data":"4377405da483b371e4a95de1d521efb2185ee422736bbcf1fa7efcd584c0ef4a"} Mar 12 00:20:02 crc kubenswrapper[4948]: I0312 00:20:02.708243 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-85fd558b5-jvl2f" event={"ID":"406df0e7-9930-440b-9023-5369cdac443b","Type":"ContainerStarted","Data":"50b11efc2957fabfbfa5d1e334373e40daa8dd28f92ff9c27084e21f4ab1a916"} Mar 12 00:20:02 crc kubenswrapper[4948]: I0312 00:20:02.709227 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29554580-ghvkh" event={"ID":"4696b53e-388b-457f-9dc9-93811078ccc4","Type":"ContainerStarted","Data":"a7440a5c939651f60fea120b345322a008a40a836b1b833507aac8feb7decb6a"} Mar 12 00:20:02 crc kubenswrapper[4948]: I0312 00:20:02.710971 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5bf474d74f-swd7r" event={"ID":"f7223c5e-0d07-4513-a05a-3d504647724c","Type":"ContainerStarted","Data":"3e47585250c06fc7ea1987a8b0311257eb4dfaca430fd13b6402db0264d18c97"} Mar 12 00:20:02 crc kubenswrapper[4948]: I0312 00:20:02.713042 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-cvw7f" event={"ID":"b5f03616-ebd8-4698-91f7-ca2419301e08","Type":"ContainerStarted","Data":"099e44f07f0e64ba81d9fe9f62c4309d7ab84dc377d5cd157144e2a3e34bc118"} Mar 12 00:20:02 crc kubenswrapper[4948]: I0312 00:20:02.716641 4948 generic.go:334] "Generic (PLEG): container finished" podID="3aa75b6b-35be-412c-8fa5-1e16b61a40fc" containerID="eb8a9c314320f157c8ab769b3c6ef55d19e8535fbba9081094d0d1afe1fcad88" exitCode=0 Mar 12 00:20:02 crc kubenswrapper[4948]: I0312 00:20:02.716690 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5j9jsp" event={"ID":"3aa75b6b-35be-412c-8fa5-1e16b61a40fc","Type":"ContainerDied","Data":"eb8a9c314320f157c8ab769b3c6ef55d19e8535fbba9081094d0d1afe1fcad88"} Mar 12 00:20:03 crc kubenswrapper[4948]: I0312 00:20:03.724545 4948 generic.go:334] "Generic (PLEG): container finished" podID="3aa75b6b-35be-412c-8fa5-1e16b61a40fc" containerID="299db9031e8ade37b44118a07df328bbb60a49b5b48ceac324bb970a46fce5a9" exitCode=0 Mar 12 00:20:03 crc kubenswrapper[4948]: I0312 00:20:03.724579 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5j9jsp" event={"ID":"3aa75b6b-35be-412c-8fa5-1e16b61a40fc","Type":"ContainerDied","Data":"299db9031e8ade37b44118a07df328bbb60a49b5b48ceac324bb970a46fce5a9"} Mar 12 00:20:04 crc kubenswrapper[4948]: I0312 00:20:04.766608 4948 generic.go:334] "Generic (PLEG): container finished" podID="4696b53e-388b-457f-9dc9-93811078ccc4" containerID="381301100f845275e861e7f705307b29cde25017428f668242d0e84e87383c4a" exitCode=0 Mar 12 00:20:04 crc kubenswrapper[4948]: I0312 00:20:04.767059 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29554580-ghvkh" event={"ID":"4696b53e-388b-457f-9dc9-93811078ccc4","Type":"ContainerDied","Data":"381301100f845275e861e7f705307b29cde25017428f668242d0e84e87383c4a"} Mar 12 00:20:05 crc kubenswrapper[4948]: I0312 00:20:05.044478 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5j9jsp" Mar 12 00:20:05 crc kubenswrapper[4948]: I0312 00:20:05.079721 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3aa75b6b-35be-412c-8fa5-1e16b61a40fc-util\") pod \"3aa75b6b-35be-412c-8fa5-1e16b61a40fc\" (UID: \"3aa75b6b-35be-412c-8fa5-1e16b61a40fc\") " Mar 12 00:20:05 crc kubenswrapper[4948]: I0312 00:20:05.079823 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-ncfd2\" (UniqueName: \"kubernetes.io/projected/3aa75b6b-35be-412c-8fa5-1e16b61a40fc-kube-api-access-ncfd2\") pod \"3aa75b6b-35be-412c-8fa5-1e16b61a40fc\" (UID: \"3aa75b6b-35be-412c-8fa5-1e16b61a40fc\") " Mar 12 00:20:05 crc kubenswrapper[4948]: I0312 00:20:05.079929 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3aa75b6b-35be-412c-8fa5-1e16b61a40fc-bundle\") pod \"3aa75b6b-35be-412c-8fa5-1e16b61a40fc\" (UID: \"3aa75b6b-35be-412c-8fa5-1e16b61a40fc\") " Mar 12 00:20:05 crc kubenswrapper[4948]: I0312 00:20:05.082541 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3aa75b6b-35be-412c-8fa5-1e16b61a40fc-bundle" (OuterVolumeSpecName: "bundle") pod "3aa75b6b-35be-412c-8fa5-1e16b61a40fc" (UID: "3aa75b6b-35be-412c-8fa5-1e16b61a40fc"). InnerVolumeSpecName "bundle". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 00:20:05 crc kubenswrapper[4948]: I0312 00:20:05.100674 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/3aa75b6b-35be-412c-8fa5-1e16b61a40fc-util" (OuterVolumeSpecName: "util") pod "3aa75b6b-35be-412c-8fa5-1e16b61a40fc" (UID: "3aa75b6b-35be-412c-8fa5-1e16b61a40fc"). InnerVolumeSpecName "util". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 00:20:05 crc kubenswrapper[4948]: I0312 00:20:05.102913 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/3aa75b6b-35be-412c-8fa5-1e16b61a40fc-kube-api-access-ncfd2" (OuterVolumeSpecName: "kube-api-access-ncfd2") pod "3aa75b6b-35be-412c-8fa5-1e16b61a40fc" (UID: "3aa75b6b-35be-412c-8fa5-1e16b61a40fc"). InnerVolumeSpecName "kube-api-access-ncfd2". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:20:05 crc kubenswrapper[4948]: I0312 00:20:05.181644 4948 reconciler_common.go:293] "Volume detached for volume \"util\" (UniqueName: \"kubernetes.io/empty-dir/3aa75b6b-35be-412c-8fa5-1e16b61a40fc-util\") on node \"crc\" DevicePath \"\"" Mar 12 00:20:05 crc kubenswrapper[4948]: I0312 00:20:05.181685 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-ncfd2\" (UniqueName: \"kubernetes.io/projected/3aa75b6b-35be-412c-8fa5-1e16b61a40fc-kube-api-access-ncfd2\") on node \"crc\" DevicePath \"\"" Mar 12 00:20:05 crc kubenswrapper[4948]: I0312 00:20:05.181699 4948 reconciler_common.go:293] "Volume detached for volume \"bundle\" (UniqueName: \"kubernetes.io/empty-dir/3aa75b6b-35be-412c-8fa5-1e16b61a40fc-bundle\") on node \"crc\" DevicePath \"\"" Mar 12 00:20:05 crc kubenswrapper[4948]: I0312 00:20:05.775406 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5j9jsp" event={"ID":"3aa75b6b-35be-412c-8fa5-1e16b61a40fc","Type":"ContainerDied","Data":"8b9bc6ddbdbdb820665ed558943e3f305b8ea05474037a63d3669ad755a9b172"} Mar 12 00:20:05 crc kubenswrapper[4948]: I0312 00:20:05.775456 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="8b9bc6ddbdbdb820665ed558943e3f305b8ea05474037a63d3669ad755a9b172" Mar 12 00:20:05 crc kubenswrapper[4948]: I0312 00:20:05.775494 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5j9jsp" Mar 12 00:20:08 crc kubenswrapper[4948]: I0312 00:20:08.655239 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/interconnect-operator-5bb49f789d-tfkll"] Mar 12 00:20:08 crc kubenswrapper[4948]: E0312 00:20:08.655677 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3aa75b6b-35be-412c-8fa5-1e16b61a40fc" containerName="extract" Mar 12 00:20:08 crc kubenswrapper[4948]: I0312 00:20:08.655689 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="3aa75b6b-35be-412c-8fa5-1e16b61a40fc" containerName="extract" Mar 12 00:20:08 crc kubenswrapper[4948]: E0312 00:20:08.655704 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3aa75b6b-35be-412c-8fa5-1e16b61a40fc" containerName="util" Mar 12 00:20:08 crc kubenswrapper[4948]: I0312 00:20:08.655710 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="3aa75b6b-35be-412c-8fa5-1e16b61a40fc" containerName="util" Mar 12 00:20:08 crc kubenswrapper[4948]: E0312 00:20:08.655718 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="3aa75b6b-35be-412c-8fa5-1e16b61a40fc" containerName="pull" Mar 12 00:20:08 crc kubenswrapper[4948]: I0312 00:20:08.655724 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="3aa75b6b-35be-412c-8fa5-1e16b61a40fc" containerName="pull" Mar 12 00:20:08 crc kubenswrapper[4948]: I0312 00:20:08.655817 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="3aa75b6b-35be-412c-8fa5-1e16b61a40fc" containerName="extract" Mar 12 00:20:08 crc kubenswrapper[4948]: I0312 00:20:08.656174 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/interconnect-operator-5bb49f789d-tfkll" Mar 12 00:20:08 crc kubenswrapper[4948]: I0312 00:20:08.660477 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"kube-root-ca.crt" Mar 12 00:20:08 crc kubenswrapper[4948]: I0312 00:20:08.660839 4948 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"interconnect-operator-dockercfg-r7wgp" Mar 12 00:20:08 crc kubenswrapper[4948]: I0312 00:20:08.661938 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"openshift-service-ca.crt" Mar 12 00:20:08 crc kubenswrapper[4948]: I0312 00:20:08.676208 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/interconnect-operator-5bb49f789d-tfkll"] Mar 12 00:20:08 crc kubenswrapper[4948]: I0312 00:20:08.737173 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-fxpjr\" (UniqueName: \"kubernetes.io/projected/edf23713-864e-4d07-b866-17bbba5cd7bd-kube-api-access-fxpjr\") pod \"interconnect-operator-5bb49f789d-tfkll\" (UID: \"edf23713-864e-4d07-b866-17bbba5cd7bd\") " pod="service-telemetry/interconnect-operator-5bb49f789d-tfkll" Mar 12 00:20:08 crc kubenswrapper[4948]: I0312 00:20:08.837928 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-fxpjr\" (UniqueName: \"kubernetes.io/projected/edf23713-864e-4d07-b866-17bbba5cd7bd-kube-api-access-fxpjr\") pod \"interconnect-operator-5bb49f789d-tfkll\" (UID: \"edf23713-864e-4d07-b866-17bbba5cd7bd\") " pod="service-telemetry/interconnect-operator-5bb49f789d-tfkll" Mar 12 00:20:08 crc kubenswrapper[4948]: I0312 00:20:08.856553 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-fxpjr\" (UniqueName: \"kubernetes.io/projected/edf23713-864e-4d07-b866-17bbba5cd7bd-kube-api-access-fxpjr\") pod \"interconnect-operator-5bb49f789d-tfkll\" (UID: \"edf23713-864e-4d07-b866-17bbba5cd7bd\") " pod="service-telemetry/interconnect-operator-5bb49f789d-tfkll" Mar 12 00:20:08 crc kubenswrapper[4948]: I0312 00:20:08.969208 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/interconnect-operator-5bb49f789d-tfkll" Mar 12 00:20:09 crc kubenswrapper[4948]: I0312 00:20:09.923760 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29554580-ghvkh" Mar 12 00:20:09 crc kubenswrapper[4948]: I0312 00:20:09.952067 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4wb5d\" (UniqueName: \"kubernetes.io/projected/4696b53e-388b-457f-9dc9-93811078ccc4-kube-api-access-4wb5d\") pod \"4696b53e-388b-457f-9dc9-93811078ccc4\" (UID: \"4696b53e-388b-457f-9dc9-93811078ccc4\") " Mar 12 00:20:09 crc kubenswrapper[4948]: I0312 00:20:09.955835 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/4696b53e-388b-457f-9dc9-93811078ccc4-kube-api-access-4wb5d" (OuterVolumeSpecName: "kube-api-access-4wb5d") pod "4696b53e-388b-457f-9dc9-93811078ccc4" (UID: "4696b53e-388b-457f-9dc9-93811078ccc4"). InnerVolumeSpecName "kube-api-access-4wb5d". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:20:10 crc kubenswrapper[4948]: I0312 00:20:10.053735 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4wb5d\" (UniqueName: \"kubernetes.io/projected/4696b53e-388b-457f-9dc9-93811078ccc4-kube-api-access-4wb5d\") on node \"crc\" DevicePath \"\"" Mar 12 00:20:10 crc kubenswrapper[4948]: I0312 00:20:10.815709 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29554580-ghvkh" event={"ID":"4696b53e-388b-457f-9dc9-93811078ccc4","Type":"ContainerDied","Data":"a7440a5c939651f60fea120b345322a008a40a836b1b833507aac8feb7decb6a"} Mar 12 00:20:10 crc kubenswrapper[4948]: I0312 00:20:10.815746 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="a7440a5c939651f60fea120b345322a008a40a836b1b833507aac8feb7decb6a" Mar 12 00:20:10 crc kubenswrapper[4948]: I0312 00:20:10.815790 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29554580-ghvkh" Mar 12 00:20:10 crc kubenswrapper[4948]: I0312 00:20:10.853802 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/elastic-operator-77bd5799b9-vnntm"] Mar 12 00:20:10 crc kubenswrapper[4948]: E0312 00:20:10.853995 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="4696b53e-388b-457f-9dc9-93811078ccc4" containerName="oc" Mar 12 00:20:10 crc kubenswrapper[4948]: I0312 00:20:10.854007 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="4696b53e-388b-457f-9dc9-93811078ccc4" containerName="oc" Mar 12 00:20:10 crc kubenswrapper[4948]: I0312 00:20:10.854103 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="4696b53e-388b-457f-9dc9-93811078ccc4" containerName="oc" Mar 12 00:20:10 crc kubenswrapper[4948]: I0312 00:20:10.854467 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elastic-operator-77bd5799b9-vnntm" Mar 12 00:20:10 crc kubenswrapper[4948]: I0312 00:20:10.856708 4948 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elastic-operator-dockercfg-g5f59" Mar 12 00:20:10 crc kubenswrapper[4948]: I0312 00:20:10.858715 4948 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elastic-operator-service-cert" Mar 12 00:20:10 crc kubenswrapper[4948]: I0312 00:20:10.863457 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e47c980d-dc98-45a8-b10a-768351396c21-webhook-cert\") pod \"elastic-operator-77bd5799b9-vnntm\" (UID: \"e47c980d-dc98-45a8-b10a-768351396c21\") " pod="service-telemetry/elastic-operator-77bd5799b9-vnntm" Mar 12 00:20:10 crc kubenswrapper[4948]: I0312 00:20:10.863499 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-mjjsq\" (UniqueName: \"kubernetes.io/projected/e47c980d-dc98-45a8-b10a-768351396c21-kube-api-access-mjjsq\") pod \"elastic-operator-77bd5799b9-vnntm\" (UID: \"e47c980d-dc98-45a8-b10a-768351396c21\") " pod="service-telemetry/elastic-operator-77bd5799b9-vnntm" Mar 12 00:20:10 crc kubenswrapper[4948]: I0312 00:20:10.863547 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e47c980d-dc98-45a8-b10a-768351396c21-apiservice-cert\") pod \"elastic-operator-77bd5799b9-vnntm\" (UID: \"e47c980d-dc98-45a8-b10a-768351396c21\") " pod="service-telemetry/elastic-operator-77bd5799b9-vnntm" Mar 12 00:20:10 crc kubenswrapper[4948]: I0312 00:20:10.892154 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elastic-operator-77bd5799b9-vnntm"] Mar 12 00:20:10 crc kubenswrapper[4948]: I0312 00:20:10.963642 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29554574-5h7vf"] Mar 12 00:20:10 crc kubenswrapper[4948]: I0312 00:20:10.963889 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e47c980d-dc98-45a8-b10a-768351396c21-apiservice-cert\") pod \"elastic-operator-77bd5799b9-vnntm\" (UID: \"e47c980d-dc98-45a8-b10a-768351396c21\") " pod="service-telemetry/elastic-operator-77bd5799b9-vnntm" Mar 12 00:20:10 crc kubenswrapper[4948]: I0312 00:20:10.963944 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e47c980d-dc98-45a8-b10a-768351396c21-webhook-cert\") pod \"elastic-operator-77bd5799b9-vnntm\" (UID: \"e47c980d-dc98-45a8-b10a-768351396c21\") " pod="service-telemetry/elastic-operator-77bd5799b9-vnntm" Mar 12 00:20:10 crc kubenswrapper[4948]: I0312 00:20:10.963964 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-mjjsq\" (UniqueName: \"kubernetes.io/projected/e47c980d-dc98-45a8-b10a-768351396c21-kube-api-access-mjjsq\") pod \"elastic-operator-77bd5799b9-vnntm\" (UID: \"e47c980d-dc98-45a8-b10a-768351396c21\") " pod="service-telemetry/elastic-operator-77bd5799b9-vnntm" Mar 12 00:20:10 crc kubenswrapper[4948]: I0312 00:20:10.967664 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"apiservice-cert\" (UniqueName: \"kubernetes.io/secret/e47c980d-dc98-45a8-b10a-768351396c21-apiservice-cert\") pod \"elastic-operator-77bd5799b9-vnntm\" (UID: \"e47c980d-dc98-45a8-b10a-768351396c21\") " pod="service-telemetry/elastic-operator-77bd5799b9-vnntm" Mar 12 00:20:10 crc kubenswrapper[4948]: I0312 00:20:10.968144 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"webhook-cert\" (UniqueName: \"kubernetes.io/secret/e47c980d-dc98-45a8-b10a-768351396c21-webhook-cert\") pod \"elastic-operator-77bd5799b9-vnntm\" (UID: \"e47c980d-dc98-45a8-b10a-768351396c21\") " pod="service-telemetry/elastic-operator-77bd5799b9-vnntm" Mar 12 00:20:10 crc kubenswrapper[4948]: I0312 00:20:10.974018 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29554574-5h7vf"] Mar 12 00:20:10 crc kubenswrapper[4948]: I0312 00:20:10.980948 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-mjjsq\" (UniqueName: \"kubernetes.io/projected/e47c980d-dc98-45a8-b10a-768351396c21-kube-api-access-mjjsq\") pod \"elastic-operator-77bd5799b9-vnntm\" (UID: \"e47c980d-dc98-45a8-b10a-768351396c21\") " pod="service-telemetry/elastic-operator-77bd5799b9-vnntm" Mar 12 00:20:11 crc kubenswrapper[4948]: I0312 00:20:11.168735 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elastic-operator-77bd5799b9-vnntm" Mar 12 00:20:11 crc kubenswrapper[4948]: I0312 00:20:11.322511 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="f9a50fc1-6ae2-4bb0-91f5-2e51b77d1b36" path="/var/lib/kubelet/pods/f9a50fc1-6ae2-4bb0-91f5-2e51b77d1b36/volumes" Mar 12 00:20:12 crc kubenswrapper[4948]: I0312 00:20:12.914576 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/interconnect-operator-5bb49f789d-tfkll"] Mar 12 00:20:12 crc kubenswrapper[4948]: I0312 00:20:12.990601 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elastic-operator-77bd5799b9-vnntm"] Mar 12 00:20:12 crc kubenswrapper[4948]: W0312 00:20:12.990842 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pode47c980d_dc98_45a8_b10a_768351396c21.slice/crio-abfffc846b4a6c559ef266aac279a19f6474db4a0176a5a5789c96c0dc930085 WatchSource:0}: Error finding container abfffc846b4a6c559ef266aac279a19f6474db4a0176a5a5789c96c0dc930085: Status 404 returned error can't find the container with id abfffc846b4a6c559ef266aac279a19f6474db4a0176a5a5789c96c0dc930085 Mar 12 00:20:13 crc kubenswrapper[4948]: I0312 00:20:13.837899 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-cvw7f" event={"ID":"b5f03616-ebd8-4698-91f7-ca2419301e08","Type":"ContainerStarted","Data":"2611a9b3dffc0ea597ab7377bae1da9f0f29b8bbd669fe14235412e161f8c394"} Mar 12 00:20:13 crc kubenswrapper[4948]: I0312 00:20:13.839492 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/interconnect-operator-5bb49f789d-tfkll" event={"ID":"edf23713-864e-4d07-b866-17bbba5cd7bd","Type":"ContainerStarted","Data":"94e3b463d0304a0225cf5a2954e2c87e622967ba0da96873be051d2a0dc8390e"} Mar 12 00:20:13 crc kubenswrapper[4948]: I0312 00:20:13.841282 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/observability-operator-59bdc8b94-bkzp4" event={"ID":"bf64ec59-b44a-407b-8d10-f1f9bd91de76","Type":"ContainerStarted","Data":"f795f695916a8c22d5ac0c7e88138a04b5a418ae1d400e88b75aaacfc7559d5b"} Mar 12 00:20:13 crc kubenswrapper[4948]: I0312 00:20:13.841484 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/observability-operator-59bdc8b94-bkzp4" Mar 12 00:20:13 crc kubenswrapper[4948]: I0312 00:20:13.843259 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/observability-operator-59bdc8b94-bkzp4" Mar 12 00:20:13 crc kubenswrapper[4948]: I0312 00:20:13.844368 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-85fd558b5-d55ck" event={"ID":"06125956-526b-468b-b609-9bb13c834499","Type":"ContainerStarted","Data":"117962a7d6404d09fc14d873a4d07a83b8b1d5d52ace22c60f92b2dc2e55da55"} Mar 12 00:20:13 crc kubenswrapper[4948]: I0312 00:20:13.845770 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/obo-prometheus-operator-admission-webhook-85fd558b5-jvl2f" event={"ID":"406df0e7-9930-440b-9023-5369cdac443b","Type":"ContainerStarted","Data":"2c160937267c6d17b1f0baaa5aafd704b2c02e1c79a379e2c6fb2e4a73d44764"} Mar 12 00:20:13 crc kubenswrapper[4948]: I0312 00:20:13.847234 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elastic-operator-77bd5799b9-vnntm" event={"ID":"e47c980d-dc98-45a8-b10a-768351396c21","Type":"ContainerStarted","Data":"abfffc846b4a6c559ef266aac279a19f6474db4a0176a5a5789c96c0dc930085"} Mar 12 00:20:13 crc kubenswrapper[4948]: I0312 00:20:13.848408 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-operators/perses-operator-5bf474d74f-swd7r" event={"ID":"f7223c5e-0d07-4513-a05a-3d504647724c","Type":"ContainerStarted","Data":"04a573f1ea8bda995f59c80b81e9291c1e5dd2758608fb5f76215eb338d7dbf9"} Mar 12 00:20:13 crc kubenswrapper[4948]: I0312 00:20:13.848573 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-operators/perses-operator-5bf474d74f-swd7r" Mar 12 00:20:13 crc kubenswrapper[4948]: I0312 00:20:13.859845 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-68bc856cb9-cvw7f" podStartSLOduration=2.319989028 podStartE2EDuration="12.859823215s" podCreationTimestamp="2026-03-12 00:20:01 +0000 UTC" firstStartedPulling="2026-03-12 00:20:02.080830746 +0000 UTC m=+801.536434484" lastFinishedPulling="2026-03-12 00:20:12.620664913 +0000 UTC m=+812.076268671" observedRunningTime="2026-03-12 00:20:13.849858384 +0000 UTC m=+813.305462122" watchObservedRunningTime="2026-03-12 00:20:13.859823215 +0000 UTC m=+813.315426953" Mar 12 00:20:13 crc kubenswrapper[4948]: I0312 00:20:13.868405 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/perses-operator-5bf474d74f-swd7r" podStartSLOduration=2.508074889 podStartE2EDuration="12.868387347s" podCreationTimestamp="2026-03-12 00:20:01 +0000 UTC" firstStartedPulling="2026-03-12 00:20:02.278103496 +0000 UTC m=+801.733707234" lastFinishedPulling="2026-03-12 00:20:12.638415934 +0000 UTC m=+812.094019692" observedRunningTime="2026-03-12 00:20:13.86591058 +0000 UTC m=+813.321514318" watchObservedRunningTime="2026-03-12 00:20:13.868387347 +0000 UTC m=+813.323991085" Mar 12 00:20:13 crc kubenswrapper[4948]: I0312 00:20:13.887021 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/observability-operator-59bdc8b94-bkzp4" podStartSLOduration=2.363420586 podStartE2EDuration="12.887001562s" podCreationTimestamp="2026-03-12 00:20:01 +0000 UTC" firstStartedPulling="2026-03-12 00:20:02.172256675 +0000 UTC m=+801.627860403" lastFinishedPulling="2026-03-12 00:20:12.695837641 +0000 UTC m=+812.151441379" observedRunningTime="2026-03-12 00:20:13.885203383 +0000 UTC m=+813.340807121" watchObservedRunningTime="2026-03-12 00:20:13.887001562 +0000 UTC m=+813.342605310" Mar 12 00:20:13 crc kubenswrapper[4948]: I0312 00:20:13.949869 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-85fd558b5-d55ck" podStartSLOduration=2.31785201 podStartE2EDuration="12.949847416s" podCreationTimestamp="2026-03-12 00:20:01 +0000 UTC" firstStartedPulling="2026-03-12 00:20:02.079672515 +0000 UTC m=+801.535276263" lastFinishedPulling="2026-03-12 00:20:12.711667931 +0000 UTC m=+812.167271669" observedRunningTime="2026-03-12 00:20:13.946570818 +0000 UTC m=+813.402174546" watchObservedRunningTime="2026-03-12 00:20:13.949847416 +0000 UTC m=+813.405451154" Mar 12 00:20:13 crc kubenswrapper[4948]: I0312 00:20:13.977367 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-operators/obo-prometheus-operator-admission-webhook-85fd558b5-jvl2f" podStartSLOduration=2.512674054 podStartE2EDuration="12.977353092s" podCreationTimestamp="2026-03-12 00:20:01 +0000 UTC" firstStartedPulling="2026-03-12 00:20:02.173373676 +0000 UTC m=+801.628977424" lastFinishedPulling="2026-03-12 00:20:12.638052714 +0000 UTC m=+812.093656462" observedRunningTime="2026-03-12 00:20:13.97543752 +0000 UTC m=+813.431041268" watchObservedRunningTime="2026-03-12 00:20:13.977353092 +0000 UTC m=+813.432956830" Mar 12 00:20:16 crc kubenswrapper[4948]: I0312 00:20:16.881018 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elastic-operator-77bd5799b9-vnntm" event={"ID":"e47c980d-dc98-45a8-b10a-768351396c21","Type":"ContainerStarted","Data":"2c174971cff71f27d8c3403bd383366150621f88f34a85c39d0ce8a70be3bb3d"} Mar 12 00:20:16 crc kubenswrapper[4948]: I0312 00:20:16.904812 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/elastic-operator-77bd5799b9-vnntm" podStartSLOduration=3.861006347 podStartE2EDuration="6.904791124s" podCreationTimestamp="2026-03-12 00:20:10 +0000 UTC" firstStartedPulling="2026-03-12 00:20:12.99336682 +0000 UTC m=+812.448970558" lastFinishedPulling="2026-03-12 00:20:16.037151597 +0000 UTC m=+815.492755335" observedRunningTime="2026-03-12 00:20:16.899408569 +0000 UTC m=+816.355012317" watchObservedRunningTime="2026-03-12 00:20:16.904791124 +0000 UTC m=+816.360394872" Mar 12 00:20:18 crc kubenswrapper[4948]: I0312 00:20:18.719198 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-6dzwm"] Mar 12 00:20:18 crc kubenswrapper[4948]: I0312 00:20:18.727617 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-6dzwm" Mar 12 00:20:18 crc kubenswrapper[4948]: I0312 00:20:18.732542 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"kube-root-ca.crt" Mar 12 00:20:18 crc kubenswrapper[4948]: I0312 00:20:18.732836 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager-operator"/"openshift-service-ca.crt" Mar 12 00:20:18 crc kubenswrapper[4948]: I0312 00:20:18.732981 4948 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager-operator"/"cert-manager-operator-controller-manager-dockercfg-q94db" Mar 12 00:20:18 crc kubenswrapper[4948]: I0312 00:20:18.737049 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-6dzwm"] Mar 12 00:20:18 crc kubenswrapper[4948]: I0312 00:20:18.920139 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-g945d\" (UniqueName: \"kubernetes.io/projected/264675fd-de87-419e-bef5-c833bb0aedf2-kube-api-access-g945d\") pod \"cert-manager-operator-controller-manager-5586865c96-6dzwm\" (UID: \"264675fd-de87-419e-bef5-c833bb0aedf2\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-6dzwm" Mar 12 00:20:18 crc kubenswrapper[4948]: I0312 00:20:18.920195 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/264675fd-de87-419e-bef5-c833bb0aedf2-tmp\") pod \"cert-manager-operator-controller-manager-5586865c96-6dzwm\" (UID: \"264675fd-de87-419e-bef5-c833bb0aedf2\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-6dzwm" Mar 12 00:20:19 crc kubenswrapper[4948]: I0312 00:20:19.021535 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-g945d\" (UniqueName: \"kubernetes.io/projected/264675fd-de87-419e-bef5-c833bb0aedf2-kube-api-access-g945d\") pod \"cert-manager-operator-controller-manager-5586865c96-6dzwm\" (UID: \"264675fd-de87-419e-bef5-c833bb0aedf2\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-6dzwm" Mar 12 00:20:19 crc kubenswrapper[4948]: I0312 00:20:19.021609 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/264675fd-de87-419e-bef5-c833bb0aedf2-tmp\") pod \"cert-manager-operator-controller-manager-5586865c96-6dzwm\" (UID: \"264675fd-de87-419e-bef5-c833bb0aedf2\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-6dzwm" Mar 12 00:20:19 crc kubenswrapper[4948]: I0312 00:20:19.022200 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp\" (UniqueName: \"kubernetes.io/empty-dir/264675fd-de87-419e-bef5-c833bb0aedf2-tmp\") pod \"cert-manager-operator-controller-manager-5586865c96-6dzwm\" (UID: \"264675fd-de87-419e-bef5-c833bb0aedf2\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-6dzwm" Mar 12 00:20:19 crc kubenswrapper[4948]: I0312 00:20:19.040000 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-g945d\" (UniqueName: \"kubernetes.io/projected/264675fd-de87-419e-bef5-c833bb0aedf2-kube-api-access-g945d\") pod \"cert-manager-operator-controller-manager-5586865c96-6dzwm\" (UID: \"264675fd-de87-419e-bef5-c833bb0aedf2\") " pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-6dzwm" Mar 12 00:20:19 crc kubenswrapper[4948]: I0312 00:20:19.082363 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-6dzwm" Mar 12 00:20:19 crc kubenswrapper[4948]: I0312 00:20:19.645661 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Mar 12 00:20:19 crc kubenswrapper[4948]: I0312 00:20:19.647064 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elasticsearch-es-default-0" Mar 12 00:20:19 crc kubenswrapper[4948]: I0312 00:20:19.660615 4948 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-internal-users" Mar 12 00:20:19 crc kubenswrapper[4948]: I0312 00:20:19.666234 4948 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"default-dockercfg-x8tmd" Mar 12 00:20:19 crc kubenswrapper[4948]: I0312 00:20:19.666873 4948 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-xpack-file-realm" Mar 12 00:20:19 crc kubenswrapper[4948]: I0312 00:20:19.676814 4948 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-remote-ca" Mar 12 00:20:19 crc kubenswrapper[4948]: I0312 00:20:19.677129 4948 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-default-es-config" Mar 12 00:20:19 crc kubenswrapper[4948]: I0312 00:20:19.677630 4948 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-default-es-transport-certs" Mar 12 00:20:19 crc kubenswrapper[4948]: I0312 00:20:19.678047 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"elasticsearch-es-scripts" Mar 12 00:20:19 crc kubenswrapper[4948]: I0312 00:20:19.678881 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"service-telemetry"/"elasticsearch-es-unicast-hosts" Mar 12 00:20:19 crc kubenswrapper[4948]: I0312 00:20:19.679079 4948 reflector.go:368] Caches populated for *v1.Secret from object-"service-telemetry"/"elasticsearch-es-http-certs-internal" Mar 12 00:20:19 crc kubenswrapper[4948]: I0312 00:20:19.688021 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Mar 12 00:20:19 crc kubenswrapper[4948]: I0312 00:20:19.729959 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-bin-local\" (UniqueName: \"kubernetes.io/empty-dir/480695a0-04b7-4ef9-9625-10ff480f8524-elastic-internal-elasticsearch-bin-local\") pod \"elasticsearch-es-default-0\" (UID: \"480695a0-04b7-4ef9-9625-10ff480f8524\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 12 00:20:19 crc kubenswrapper[4948]: I0312 00:20:19.730039 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-scripts\" (UniqueName: \"kubernetes.io/configmap/480695a0-04b7-4ef9-9625-10ff480f8524-elastic-internal-scripts\") pod \"elasticsearch-es-default-0\" (UID: \"480695a0-04b7-4ef9-9625-10ff480f8524\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 12 00:20:19 crc kubenswrapper[4948]: I0312 00:20:19.730076 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elasticsearch-logs\" (UniqueName: \"kubernetes.io/empty-dir/480695a0-04b7-4ef9-9625-10ff480f8524-elasticsearch-logs\") pod \"elasticsearch-es-default-0\" (UID: \"480695a0-04b7-4ef9-9625-10ff480f8524\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 12 00:20:19 crc kubenswrapper[4948]: I0312 00:20:19.730113 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-probe-user\" (UniqueName: \"kubernetes.io/secret/480695a0-04b7-4ef9-9625-10ff480f8524-elastic-internal-probe-user\") pod \"elasticsearch-es-default-0\" (UID: \"480695a0-04b7-4ef9-9625-10ff480f8524\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 12 00:20:19 crc kubenswrapper[4948]: I0312 00:20:19.730142 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-config-local\" (UniqueName: \"kubernetes.io/empty-dir/480695a0-04b7-4ef9-9625-10ff480f8524-elastic-internal-elasticsearch-config-local\") pod \"elasticsearch-es-default-0\" (UID: \"480695a0-04b7-4ef9-9625-10ff480f8524\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 12 00:20:19 crc kubenswrapper[4948]: I0312 00:20:19.730170 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-http-certificates\" (UniqueName: \"kubernetes.io/secret/480695a0-04b7-4ef9-9625-10ff480f8524-elastic-internal-http-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"480695a0-04b7-4ef9-9625-10ff480f8524\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 12 00:20:19 crc kubenswrapper[4948]: I0312 00:20:19.730195 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-transport-certificates\" (UniqueName: \"kubernetes.io/secret/480695a0-04b7-4ef9-9625-10ff480f8524-elastic-internal-transport-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"480695a0-04b7-4ef9-9625-10ff480f8524\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 12 00:20:19 crc kubenswrapper[4948]: I0312 00:20:19.730221 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-plugins-local\" (UniqueName: \"kubernetes.io/empty-dir/480695a0-04b7-4ef9-9625-10ff480f8524-elastic-internal-elasticsearch-plugins-local\") pod \"elasticsearch-es-default-0\" (UID: \"480695a0-04b7-4ef9-9625-10ff480f8524\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 12 00:20:19 crc kubenswrapper[4948]: I0312 00:20:19.730243 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"tmp-volume\" (UniqueName: \"kubernetes.io/empty-dir/480695a0-04b7-4ef9-9625-10ff480f8524-tmp-volume\") pod \"elasticsearch-es-default-0\" (UID: \"480695a0-04b7-4ef9-9625-10ff480f8524\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 12 00:20:19 crc kubenswrapper[4948]: I0312 00:20:19.730265 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-xpack-file-realm\" (UniqueName: \"kubernetes.io/secret/480695a0-04b7-4ef9-9625-10ff480f8524-elastic-internal-xpack-file-realm\") pod \"elasticsearch-es-default-0\" (UID: \"480695a0-04b7-4ef9-9625-10ff480f8524\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 12 00:20:19 crc kubenswrapper[4948]: I0312 00:20:19.730314 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"downward-api\" (UniqueName: \"kubernetes.io/downward-api/480695a0-04b7-4ef9-9625-10ff480f8524-downward-api\") pod \"elasticsearch-es-default-0\" (UID: \"480695a0-04b7-4ef9-9625-10ff480f8524\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 12 00:20:19 crc kubenswrapper[4948]: I0312 00:20:19.730342 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-elasticsearch-config\" (UniqueName: \"kubernetes.io/secret/480695a0-04b7-4ef9-9625-10ff480f8524-elastic-internal-elasticsearch-config\") pod \"elasticsearch-es-default-0\" (UID: \"480695a0-04b7-4ef9-9625-10ff480f8524\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 12 00:20:19 crc kubenswrapper[4948]: I0312 00:20:19.730391 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-remote-certificate-authorities\" (UniqueName: \"kubernetes.io/secret/480695a0-04b7-4ef9-9625-10ff480f8524-elastic-internal-remote-certificate-authorities\") pod \"elasticsearch-es-default-0\" (UID: \"480695a0-04b7-4ef9-9625-10ff480f8524\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 12 00:20:19 crc kubenswrapper[4948]: I0312 00:20:19.730516 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elastic-internal-unicast-hosts\" (UniqueName: \"kubernetes.io/configmap/480695a0-04b7-4ef9-9625-10ff480f8524-elastic-internal-unicast-hosts\") pod \"elasticsearch-es-default-0\" (UID: \"480695a0-04b7-4ef9-9625-10ff480f8524\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 12 00:20:19 crc kubenswrapper[4948]: I0312 00:20:19.730591 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"elasticsearch-data\" (UniqueName: \"kubernetes.io/empty-dir/480695a0-04b7-4ef9-9625-10ff480f8524-elasticsearch-data\") pod \"elasticsearch-es-default-0\" (UID: \"480695a0-04b7-4ef9-9625-10ff480f8524\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 12 00:20:19 crc kubenswrapper[4948]: I0312 00:20:19.831826 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-unicast-hosts\" (UniqueName: \"kubernetes.io/configmap/480695a0-04b7-4ef9-9625-10ff480f8524-elastic-internal-unicast-hosts\") pod \"elasticsearch-es-default-0\" (UID: \"480695a0-04b7-4ef9-9625-10ff480f8524\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 12 00:20:19 crc kubenswrapper[4948]: I0312 00:20:19.831878 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elasticsearch-data\" (UniqueName: \"kubernetes.io/empty-dir/480695a0-04b7-4ef9-9625-10ff480f8524-elasticsearch-data\") pod \"elasticsearch-es-default-0\" (UID: \"480695a0-04b7-4ef9-9625-10ff480f8524\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 12 00:20:19 crc kubenswrapper[4948]: I0312 00:20:19.831901 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-bin-local\" (UniqueName: \"kubernetes.io/empty-dir/480695a0-04b7-4ef9-9625-10ff480f8524-elastic-internal-elasticsearch-bin-local\") pod \"elasticsearch-es-default-0\" (UID: \"480695a0-04b7-4ef9-9625-10ff480f8524\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 12 00:20:19 crc kubenswrapper[4948]: I0312 00:20:19.831927 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-scripts\" (UniqueName: \"kubernetes.io/configmap/480695a0-04b7-4ef9-9625-10ff480f8524-elastic-internal-scripts\") pod \"elasticsearch-es-default-0\" (UID: \"480695a0-04b7-4ef9-9625-10ff480f8524\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 12 00:20:19 crc kubenswrapper[4948]: I0312 00:20:19.831943 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elasticsearch-logs\" (UniqueName: \"kubernetes.io/empty-dir/480695a0-04b7-4ef9-9625-10ff480f8524-elasticsearch-logs\") pod \"elasticsearch-es-default-0\" (UID: \"480695a0-04b7-4ef9-9625-10ff480f8524\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 12 00:20:19 crc kubenswrapper[4948]: I0312 00:20:19.831963 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-probe-user\" (UniqueName: \"kubernetes.io/secret/480695a0-04b7-4ef9-9625-10ff480f8524-elastic-internal-probe-user\") pod \"elasticsearch-es-default-0\" (UID: \"480695a0-04b7-4ef9-9625-10ff480f8524\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 12 00:20:19 crc kubenswrapper[4948]: I0312 00:20:19.831982 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-config-local\" (UniqueName: \"kubernetes.io/empty-dir/480695a0-04b7-4ef9-9625-10ff480f8524-elastic-internal-elasticsearch-config-local\") pod \"elasticsearch-es-default-0\" (UID: \"480695a0-04b7-4ef9-9625-10ff480f8524\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 12 00:20:19 crc kubenswrapper[4948]: I0312 00:20:19.831999 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-http-certificates\" (UniqueName: \"kubernetes.io/secret/480695a0-04b7-4ef9-9625-10ff480f8524-elastic-internal-http-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"480695a0-04b7-4ef9-9625-10ff480f8524\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 12 00:20:19 crc kubenswrapper[4948]: I0312 00:20:19.832014 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-transport-certificates\" (UniqueName: \"kubernetes.io/secret/480695a0-04b7-4ef9-9625-10ff480f8524-elastic-internal-transport-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"480695a0-04b7-4ef9-9625-10ff480f8524\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 12 00:20:19 crc kubenswrapper[4948]: I0312 00:20:19.832030 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-plugins-local\" (UniqueName: \"kubernetes.io/empty-dir/480695a0-04b7-4ef9-9625-10ff480f8524-elastic-internal-elasticsearch-plugins-local\") pod \"elasticsearch-es-default-0\" (UID: \"480695a0-04b7-4ef9-9625-10ff480f8524\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 12 00:20:19 crc kubenswrapper[4948]: I0312 00:20:19.832044 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"tmp-volume\" (UniqueName: \"kubernetes.io/empty-dir/480695a0-04b7-4ef9-9625-10ff480f8524-tmp-volume\") pod \"elasticsearch-es-default-0\" (UID: \"480695a0-04b7-4ef9-9625-10ff480f8524\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 12 00:20:19 crc kubenswrapper[4948]: I0312 00:20:19.832059 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-xpack-file-realm\" (UniqueName: \"kubernetes.io/secret/480695a0-04b7-4ef9-9625-10ff480f8524-elastic-internal-xpack-file-realm\") pod \"elasticsearch-es-default-0\" (UID: \"480695a0-04b7-4ef9-9625-10ff480f8524\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 12 00:20:19 crc kubenswrapper[4948]: I0312 00:20:19.832079 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"downward-api\" (UniqueName: \"kubernetes.io/downward-api/480695a0-04b7-4ef9-9625-10ff480f8524-downward-api\") pod \"elasticsearch-es-default-0\" (UID: \"480695a0-04b7-4ef9-9625-10ff480f8524\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 12 00:20:19 crc kubenswrapper[4948]: I0312 00:20:19.832093 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-elasticsearch-config\" (UniqueName: \"kubernetes.io/secret/480695a0-04b7-4ef9-9625-10ff480f8524-elastic-internal-elasticsearch-config\") pod \"elasticsearch-es-default-0\" (UID: \"480695a0-04b7-4ef9-9625-10ff480f8524\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 12 00:20:19 crc kubenswrapper[4948]: I0312 00:20:19.832108 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"elastic-internal-remote-certificate-authorities\" (UniqueName: \"kubernetes.io/secret/480695a0-04b7-4ef9-9625-10ff480f8524-elastic-internal-remote-certificate-authorities\") pod \"elasticsearch-es-default-0\" (UID: \"480695a0-04b7-4ef9-9625-10ff480f8524\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 12 00:20:19 crc kubenswrapper[4948]: I0312 00:20:19.833106 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"tmp-volume\" (UniqueName: \"kubernetes.io/empty-dir/480695a0-04b7-4ef9-9625-10ff480f8524-tmp-volume\") pod \"elasticsearch-es-default-0\" (UID: \"480695a0-04b7-4ef9-9625-10ff480f8524\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 12 00:20:19 crc kubenswrapper[4948]: I0312 00:20:19.833415 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-unicast-hosts\" (UniqueName: \"kubernetes.io/configmap/480695a0-04b7-4ef9-9625-10ff480f8524-elastic-internal-unicast-hosts\") pod \"elasticsearch-es-default-0\" (UID: \"480695a0-04b7-4ef9-9625-10ff480f8524\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 12 00:20:19 crc kubenswrapper[4948]: I0312 00:20:19.833717 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-plugins-local\" (UniqueName: \"kubernetes.io/empty-dir/480695a0-04b7-4ef9-9625-10ff480f8524-elastic-internal-elasticsearch-plugins-local\") pod \"elasticsearch-es-default-0\" (UID: \"480695a0-04b7-4ef9-9625-10ff480f8524\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 12 00:20:19 crc kubenswrapper[4948]: I0312 00:20:19.837253 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elasticsearch-logs\" (UniqueName: \"kubernetes.io/empty-dir/480695a0-04b7-4ef9-9625-10ff480f8524-elasticsearch-logs\") pod \"elasticsearch-es-default-0\" (UID: \"480695a0-04b7-4ef9-9625-10ff480f8524\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 12 00:20:19 crc kubenswrapper[4948]: I0312 00:20:19.837718 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-scripts\" (UniqueName: \"kubernetes.io/configmap/480695a0-04b7-4ef9-9625-10ff480f8524-elastic-internal-scripts\") pod \"elasticsearch-es-default-0\" (UID: \"480695a0-04b7-4ef9-9625-10ff480f8524\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 12 00:20:19 crc kubenswrapper[4948]: I0312 00:20:19.838072 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-config-local\" (UniqueName: \"kubernetes.io/empty-dir/480695a0-04b7-4ef9-9625-10ff480f8524-elastic-internal-elasticsearch-config-local\") pod \"elasticsearch-es-default-0\" (UID: \"480695a0-04b7-4ef9-9625-10ff480f8524\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 12 00:20:19 crc kubenswrapper[4948]: I0312 00:20:19.838122 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-remote-certificate-authorities\" (UniqueName: \"kubernetes.io/secret/480695a0-04b7-4ef9-9625-10ff480f8524-elastic-internal-remote-certificate-authorities\") pod \"elasticsearch-es-default-0\" (UID: \"480695a0-04b7-4ef9-9625-10ff480f8524\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 12 00:20:19 crc kubenswrapper[4948]: I0312 00:20:19.843177 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elasticsearch-data\" (UniqueName: \"kubernetes.io/empty-dir/480695a0-04b7-4ef9-9625-10ff480f8524-elasticsearch-data\") pod \"elasticsearch-es-default-0\" (UID: \"480695a0-04b7-4ef9-9625-10ff480f8524\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 12 00:20:19 crc kubenswrapper[4948]: I0312 00:20:19.848574 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-bin-local\" (UniqueName: \"kubernetes.io/empty-dir/480695a0-04b7-4ef9-9625-10ff480f8524-elastic-internal-elasticsearch-bin-local\") pod \"elasticsearch-es-default-0\" (UID: \"480695a0-04b7-4ef9-9625-10ff480f8524\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 12 00:20:19 crc kubenswrapper[4948]: I0312 00:20:19.853954 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"downward-api\" (UniqueName: \"kubernetes.io/downward-api/480695a0-04b7-4ef9-9625-10ff480f8524-downward-api\") pod \"elasticsearch-es-default-0\" (UID: \"480695a0-04b7-4ef9-9625-10ff480f8524\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 12 00:20:19 crc kubenswrapper[4948]: I0312 00:20:19.854159 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-xpack-file-realm\" (UniqueName: \"kubernetes.io/secret/480695a0-04b7-4ef9-9625-10ff480f8524-elastic-internal-xpack-file-realm\") pod \"elasticsearch-es-default-0\" (UID: \"480695a0-04b7-4ef9-9625-10ff480f8524\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 12 00:20:19 crc kubenswrapper[4948]: I0312 00:20:19.854237 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-transport-certificates\" (UniqueName: \"kubernetes.io/secret/480695a0-04b7-4ef9-9625-10ff480f8524-elastic-internal-transport-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"480695a0-04b7-4ef9-9625-10ff480f8524\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 12 00:20:19 crc kubenswrapper[4948]: I0312 00:20:19.854593 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-elasticsearch-config\" (UniqueName: \"kubernetes.io/secret/480695a0-04b7-4ef9-9625-10ff480f8524-elastic-internal-elasticsearch-config\") pod \"elasticsearch-es-default-0\" (UID: \"480695a0-04b7-4ef9-9625-10ff480f8524\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 12 00:20:19 crc kubenswrapper[4948]: I0312 00:20:19.854697 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-probe-user\" (UniqueName: \"kubernetes.io/secret/480695a0-04b7-4ef9-9625-10ff480f8524-elastic-internal-probe-user\") pod \"elasticsearch-es-default-0\" (UID: \"480695a0-04b7-4ef9-9625-10ff480f8524\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 12 00:20:19 crc kubenswrapper[4948]: I0312 00:20:19.854711 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"elastic-internal-http-certificates\" (UniqueName: \"kubernetes.io/secret/480695a0-04b7-4ef9-9625-10ff480f8524-elastic-internal-http-certificates\") pod \"elasticsearch-es-default-0\" (UID: \"480695a0-04b7-4ef9-9625-10ff480f8524\") " pod="service-telemetry/elasticsearch-es-default-0" Mar 12 00:20:19 crc kubenswrapper[4948]: I0312 00:20:19.964710 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="service-telemetry/elasticsearch-es-default-0" Mar 12 00:20:22 crc kubenswrapper[4948]: I0312 00:20:22.050065 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-6dzwm"] Mar 12 00:20:22 crc kubenswrapper[4948]: W0312 00:20:22.060424 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-burstable.slice/kubepods-burstable-pod264675fd_de87_419e_bef5_c833bb0aedf2.slice/crio-9fa7247cbc90404c0eab53d0975e3af9b2652111829cea0fd22b79d69f8a2099 WatchSource:0}: Error finding container 9fa7247cbc90404c0eab53d0975e3af9b2652111829cea0fd22b79d69f8a2099: Status 404 returned error can't find the container with id 9fa7247cbc90404c0eab53d0975e3af9b2652111829cea0fd22b79d69f8a2099 Mar 12 00:20:22 crc kubenswrapper[4948]: I0312 00:20:22.066824 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-operators/perses-operator-5bf474d74f-swd7r" Mar 12 00:20:22 crc kubenswrapper[4948]: I0312 00:20:22.201378 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Mar 12 00:20:22 crc kubenswrapper[4948]: I0312 00:20:22.925215 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/interconnect-operator-5bb49f789d-tfkll" event={"ID":"edf23713-864e-4d07-b866-17bbba5cd7bd","Type":"ContainerStarted","Data":"e1ee8dfe8e0ba003526e56097493759191db5ea30ad1d951c5f4e74339557024"} Mar 12 00:20:22 crc kubenswrapper[4948]: I0312 00:20:22.926462 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"480695a0-04b7-4ef9-9625-10ff480f8524","Type":"ContainerStarted","Data":"6687d494ab9ec0f80518f601866fc7c652aeb27b2432ae02b53a65310513b737"} Mar 12 00:20:22 crc kubenswrapper[4948]: I0312 00:20:22.927230 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-6dzwm" event={"ID":"264675fd-de87-419e-bef5-c833bb0aedf2","Type":"ContainerStarted","Data":"9fa7247cbc90404c0eab53d0975e3af9b2652111829cea0fd22b79d69f8a2099"} Mar 12 00:20:22 crc kubenswrapper[4948]: I0312 00:20:22.955529 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/interconnect-operator-5bb49f789d-tfkll" podStartSLOduration=6.03927205 podStartE2EDuration="14.95550953s" podCreationTimestamp="2026-03-12 00:20:08 +0000 UTC" firstStartedPulling="2026-03-12 00:20:12.922028805 +0000 UTC m=+812.377632563" lastFinishedPulling="2026-03-12 00:20:21.838266305 +0000 UTC m=+821.293870043" observedRunningTime="2026-03-12 00:20:22.953041043 +0000 UTC m=+822.408644781" watchObservedRunningTime="2026-03-12 00:20:22.95550953 +0000 UTC m=+822.411113278" Mar 12 00:20:29 crc kubenswrapper[4948]: I0312 00:20:29.982104 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-6dzwm" event={"ID":"264675fd-de87-419e-bef5-c833bb0aedf2","Type":"ContainerStarted","Data":"b26fcc0189623fc99444e7c15019338a77b2f43aaa6f8d18b683a575343f7de4"} Mar 12 00:20:30 crc kubenswrapper[4948]: I0312 00:20:30.002918 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager-operator/cert-manager-operator-controller-manager-5586865c96-6dzwm" podStartSLOduration=4.361837383 podStartE2EDuration="12.002897123s" podCreationTimestamp="2026-03-12 00:20:18 +0000 UTC" firstStartedPulling="2026-03-12 00:20:22.061979312 +0000 UTC m=+821.517583050" lastFinishedPulling="2026-03-12 00:20:29.703039052 +0000 UTC m=+829.158642790" observedRunningTime="2026-03-12 00:20:30.00165726 +0000 UTC m=+829.457261018" watchObservedRunningTime="2026-03-12 00:20:30.002897123 +0000 UTC m=+829.458500861" Mar 12 00:20:31 crc kubenswrapper[4948]: I0312 00:20:31.229166 4948 dynamic_cafile_content.go:123] "Loaded a new CA Bundle and Verifier" name="client-ca-bundle::/etc/kubernetes/kubelet-ca.crt" Mar 12 00:20:36 crc kubenswrapper[4948]: I0312 00:20:36.265253 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-cainjector-5545bd876-7bjwb"] Mar 12 00:20:36 crc kubenswrapper[4948]: I0312 00:20:36.266353 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-5545bd876-7bjwb" Mar 12 00:20:36 crc kubenswrapper[4948]: I0312 00:20:36.268977 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"openshift-service-ca.crt" Mar 12 00:20:36 crc kubenswrapper[4948]: I0312 00:20:36.269502 4948 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-cainjector-dockercfg-dsk99" Mar 12 00:20:36 crc kubenswrapper[4948]: I0312 00:20:36.272319 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"cert-manager"/"kube-root-ca.crt" Mar 12 00:20:36 crc kubenswrapper[4948]: I0312 00:20:36.277031 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-5545bd876-7bjwb"] Mar 12 00:20:36 crc kubenswrapper[4948]: I0312 00:20:36.395971 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-vz69c\" (UniqueName: \"kubernetes.io/projected/dcc3f47a-6e31-417d-ab68-f4a2d50b4556-kube-api-access-vz69c\") pod \"cert-manager-cainjector-5545bd876-7bjwb\" (UID: \"dcc3f47a-6e31-417d-ab68-f4a2d50b4556\") " pod="cert-manager/cert-manager-cainjector-5545bd876-7bjwb" Mar 12 00:20:36 crc kubenswrapper[4948]: I0312 00:20:36.396023 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/dcc3f47a-6e31-417d-ab68-f4a2d50b4556-bound-sa-token\") pod \"cert-manager-cainjector-5545bd876-7bjwb\" (UID: \"dcc3f47a-6e31-417d-ab68-f4a2d50b4556\") " pod="cert-manager/cert-manager-cainjector-5545bd876-7bjwb" Mar 12 00:20:36 crc kubenswrapper[4948]: I0312 00:20:36.497572 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-vz69c\" (UniqueName: \"kubernetes.io/projected/dcc3f47a-6e31-417d-ab68-f4a2d50b4556-kube-api-access-vz69c\") pod \"cert-manager-cainjector-5545bd876-7bjwb\" (UID: \"dcc3f47a-6e31-417d-ab68-f4a2d50b4556\") " pod="cert-manager/cert-manager-cainjector-5545bd876-7bjwb" Mar 12 00:20:36 crc kubenswrapper[4948]: I0312 00:20:36.497845 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/dcc3f47a-6e31-417d-ab68-f4a2d50b4556-bound-sa-token\") pod \"cert-manager-cainjector-5545bd876-7bjwb\" (UID: \"dcc3f47a-6e31-417d-ab68-f4a2d50b4556\") " pod="cert-manager/cert-manager-cainjector-5545bd876-7bjwb" Mar 12 00:20:36 crc kubenswrapper[4948]: I0312 00:20:36.515999 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/dcc3f47a-6e31-417d-ab68-f4a2d50b4556-bound-sa-token\") pod \"cert-manager-cainjector-5545bd876-7bjwb\" (UID: \"dcc3f47a-6e31-417d-ab68-f4a2d50b4556\") " pod="cert-manager/cert-manager-cainjector-5545bd876-7bjwb" Mar 12 00:20:36 crc kubenswrapper[4948]: I0312 00:20:36.516197 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-vz69c\" (UniqueName: \"kubernetes.io/projected/dcc3f47a-6e31-417d-ab68-f4a2d50b4556-kube-api-access-vz69c\") pod \"cert-manager-cainjector-5545bd876-7bjwb\" (UID: \"dcc3f47a-6e31-417d-ab68-f4a2d50b4556\") " pod="cert-manager/cert-manager-cainjector-5545bd876-7bjwb" Mar 12 00:20:36 crc kubenswrapper[4948]: I0312 00:20:36.580773 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-cainjector-5545bd876-7bjwb" Mar 12 00:20:37 crc kubenswrapper[4948]: I0312 00:20:37.881420 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-cainjector-5545bd876-7bjwb"] Mar 12 00:20:37 crc kubenswrapper[4948]: W0312 00:20:37.885946 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-poddcc3f47a_6e31_417d_ab68_f4a2d50b4556.slice/crio-ba89ed0632dd10476ac2a347736494cd3f454be611a865aedd3db6ff39f580ff WatchSource:0}: Error finding container ba89ed0632dd10476ac2a347736494cd3f454be611a865aedd3db6ff39f580ff: Status 404 returned error can't find the container with id ba89ed0632dd10476ac2a347736494cd3f454be611a865aedd3db6ff39f580ff Mar 12 00:20:38 crc kubenswrapper[4948]: I0312 00:20:38.036670 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-5545bd876-7bjwb" event={"ID":"dcc3f47a-6e31-417d-ab68-f4a2d50b4556","Type":"ContainerStarted","Data":"ba89ed0632dd10476ac2a347736494cd3f454be611a865aedd3db6ff39f580ff"} Mar 12 00:20:38 crc kubenswrapper[4948]: I0312 00:20:38.038779 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"480695a0-04b7-4ef9-9625-10ff480f8524","Type":"ContainerStarted","Data":"101b4650848565f81c830c2107a8e07c5692982283413605acc96014d6e7883a"} Mar 12 00:20:38 crc kubenswrapper[4948]: I0312 00:20:38.261043 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Mar 12 00:20:38 crc kubenswrapper[4948]: I0312 00:20:38.303447 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["service-telemetry/elasticsearch-es-default-0"] Mar 12 00:20:39 crc kubenswrapper[4948]: I0312 00:20:39.307224 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-webhook-6888856db4-5qkhc"] Mar 12 00:20:39 crc kubenswrapper[4948]: I0312 00:20:39.308478 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-6888856db4-5qkhc" Mar 12 00:20:39 crc kubenswrapper[4948]: I0312 00:20:39.311015 4948 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-webhook-dockercfg-mw2q9" Mar 12 00:20:39 crc kubenswrapper[4948]: I0312 00:20:39.325252 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-6888856db4-5qkhc"] Mar 12 00:20:39 crc kubenswrapper[4948]: I0312 00:20:39.437024 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/50d01335-9cbd-4637-9657-27936e2265b1-bound-sa-token\") pod \"cert-manager-webhook-6888856db4-5qkhc\" (UID: \"50d01335-9cbd-4637-9657-27936e2265b1\") " pod="cert-manager/cert-manager-webhook-6888856db4-5qkhc" Mar 12 00:20:39 crc kubenswrapper[4948]: I0312 00:20:39.437209 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-xdsnr\" (UniqueName: \"kubernetes.io/projected/50d01335-9cbd-4637-9657-27936e2265b1-kube-api-access-xdsnr\") pod \"cert-manager-webhook-6888856db4-5qkhc\" (UID: \"50d01335-9cbd-4637-9657-27936e2265b1\") " pod="cert-manager/cert-manager-webhook-6888856db4-5qkhc" Mar 12 00:20:39 crc kubenswrapper[4948]: I0312 00:20:39.538826 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/50d01335-9cbd-4637-9657-27936e2265b1-bound-sa-token\") pod \"cert-manager-webhook-6888856db4-5qkhc\" (UID: \"50d01335-9cbd-4637-9657-27936e2265b1\") " pod="cert-manager/cert-manager-webhook-6888856db4-5qkhc" Mar 12 00:20:39 crc kubenswrapper[4948]: I0312 00:20:39.538912 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-xdsnr\" (UniqueName: \"kubernetes.io/projected/50d01335-9cbd-4637-9657-27936e2265b1-kube-api-access-xdsnr\") pod \"cert-manager-webhook-6888856db4-5qkhc\" (UID: \"50d01335-9cbd-4637-9657-27936e2265b1\") " pod="cert-manager/cert-manager-webhook-6888856db4-5qkhc" Mar 12 00:20:39 crc kubenswrapper[4948]: I0312 00:20:39.574411 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/50d01335-9cbd-4637-9657-27936e2265b1-bound-sa-token\") pod \"cert-manager-webhook-6888856db4-5qkhc\" (UID: \"50d01335-9cbd-4637-9657-27936e2265b1\") " pod="cert-manager/cert-manager-webhook-6888856db4-5qkhc" Mar 12 00:20:39 crc kubenswrapper[4948]: I0312 00:20:39.607163 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-xdsnr\" (UniqueName: \"kubernetes.io/projected/50d01335-9cbd-4637-9657-27936e2265b1-kube-api-access-xdsnr\") pod \"cert-manager-webhook-6888856db4-5qkhc\" (UID: \"50d01335-9cbd-4637-9657-27936e2265b1\") " pod="cert-manager/cert-manager-webhook-6888856db4-5qkhc" Mar 12 00:20:39 crc kubenswrapper[4948]: I0312 00:20:39.630998 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-webhook-6888856db4-5qkhc" Mar 12 00:20:40 crc kubenswrapper[4948]: I0312 00:20:40.055603 4948 generic.go:334] "Generic (PLEG): container finished" podID="480695a0-04b7-4ef9-9625-10ff480f8524" containerID="101b4650848565f81c830c2107a8e07c5692982283413605acc96014d6e7883a" exitCode=0 Mar 12 00:20:40 crc kubenswrapper[4948]: I0312 00:20:40.055686 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"480695a0-04b7-4ef9-9625-10ff480f8524","Type":"ContainerDied","Data":"101b4650848565f81c830c2107a8e07c5692982283413605acc96014d6e7883a"} Mar 12 00:20:40 crc kubenswrapper[4948]: I0312 00:20:40.086133 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-webhook-6888856db4-5qkhc"] Mar 12 00:20:40 crc kubenswrapper[4948]: W0312 00:20:40.093251 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod50d01335_9cbd_4637_9657_27936e2265b1.slice/crio-e2381ab3d35c4ae8d09f1d007e198a47722980fe95208fff5cb5fe9c5e753e54 WatchSource:0}: Error finding container e2381ab3d35c4ae8d09f1d007e198a47722980fe95208fff5cb5fe9c5e753e54: Status 404 returned error can't find the container with id e2381ab3d35c4ae8d09f1d007e198a47722980fe95208fff5cb5fe9c5e753e54 Mar 12 00:20:41 crc kubenswrapper[4948]: I0312 00:20:41.067782 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-6888856db4-5qkhc" event={"ID":"50d01335-9cbd-4637-9657-27936e2265b1","Type":"ContainerStarted","Data":"e2381ab3d35c4ae8d09f1d007e198a47722980fe95208fff5cb5fe9c5e753e54"} Mar 12 00:20:44 crc kubenswrapper[4948]: I0312 00:20:44.785406 4948 patch_prober.go:28] interesting pod/machine-config-daemon-m4xwc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 00:20:44 crc kubenswrapper[4948]: I0312 00:20:44.786951 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" podUID="7e483c2b-08f0-4e92-8e4a-b7281f30af3e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 00:20:45 crc kubenswrapper[4948]: I0312 00:20:45.100555 4948 generic.go:334] "Generic (PLEG): container finished" podID="480695a0-04b7-4ef9-9625-10ff480f8524" containerID="f06f766eee21751aeca3ff3cd65b6cbc910022d24212fe4d4fe0f1b3e26a4a34" exitCode=0 Mar 12 00:20:45 crc kubenswrapper[4948]: I0312 00:20:45.100598 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"480695a0-04b7-4ef9-9625-10ff480f8524","Type":"ContainerDied","Data":"f06f766eee21751aeca3ff3cd65b6cbc910022d24212fe4d4fe0f1b3e26a4a34"} Mar 12 00:20:46 crc kubenswrapper[4948]: I0312 00:20:46.111295 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="service-telemetry/elasticsearch-es-default-0" event={"ID":"480695a0-04b7-4ef9-9625-10ff480f8524","Type":"ContainerStarted","Data":"fdeedf070a01e6671e3b6fa4ec913edb7749f164e6272694e54e9d338e5a245a"} Mar 12 00:20:46 crc kubenswrapper[4948]: I0312 00:20:46.113493 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="service-telemetry/elasticsearch-es-default-0" Mar 12 00:20:46 crc kubenswrapper[4948]: I0312 00:20:46.156370 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="service-telemetry/elasticsearch-es-default-0" podStartSLOduration=11.787056431 podStartE2EDuration="27.156351487s" podCreationTimestamp="2026-03-12 00:20:19 +0000 UTC" firstStartedPulling="2026-03-12 00:20:22.197462205 +0000 UTC m=+821.653065953" lastFinishedPulling="2026-03-12 00:20:37.566757251 +0000 UTC m=+837.022361009" observedRunningTime="2026-03-12 00:20:46.153585012 +0000 UTC m=+845.609188750" watchObservedRunningTime="2026-03-12 00:20:46.156351487 +0000 UTC m=+845.611955235" Mar 12 00:20:48 crc kubenswrapper[4948]: I0312 00:20:48.122412 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-cainjector-5545bd876-7bjwb" event={"ID":"dcc3f47a-6e31-417d-ab68-f4a2d50b4556","Type":"ContainerStarted","Data":"cb11b0be089e859a383abeca305fc26bfdb94fb495a94d34738598f5ff5ec821"} Mar 12 00:20:48 crc kubenswrapper[4948]: I0312 00:20:48.124526 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-webhook-6888856db4-5qkhc" event={"ID":"50d01335-9cbd-4637-9657-27936e2265b1","Type":"ContainerStarted","Data":"651bcc93cea52ea61b7e49f1c9c034d3832d409d4a670c215bc9140be575a91c"} Mar 12 00:20:48 crc kubenswrapper[4948]: I0312 00:20:48.124935 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="cert-manager/cert-manager-webhook-6888856db4-5qkhc" Mar 12 00:20:48 crc kubenswrapper[4948]: I0312 00:20:48.145677 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-cainjector-5545bd876-7bjwb" podStartSLOduration=2.258432864 podStartE2EDuration="12.14566159s" podCreationTimestamp="2026-03-12 00:20:36 +0000 UTC" firstStartedPulling="2026-03-12 00:20:37.888960179 +0000 UTC m=+837.344563937" lastFinishedPulling="2026-03-12 00:20:47.776188925 +0000 UTC m=+847.231792663" observedRunningTime="2026-03-12 00:20:48.141882317 +0000 UTC m=+847.597486055" watchObservedRunningTime="2026-03-12 00:20:48.14566159 +0000 UTC m=+847.601265318" Mar 12 00:20:48 crc kubenswrapper[4948]: I0312 00:20:48.174492 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-webhook-6888856db4-5qkhc" podStartSLOduration=1.470297972 podStartE2EDuration="9.174475541s" podCreationTimestamp="2026-03-12 00:20:39 +0000 UTC" firstStartedPulling="2026-03-12 00:20:40.095514423 +0000 UTC m=+839.551118181" lastFinishedPulling="2026-03-12 00:20:47.799692012 +0000 UTC m=+847.255295750" observedRunningTime="2026-03-12 00:20:48.173244808 +0000 UTC m=+847.628848546" watchObservedRunningTime="2026-03-12 00:20:48.174475541 +0000 UTC m=+847.630079279" Mar 12 00:20:52 crc kubenswrapper[4948]: I0312 00:20:52.660898 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["cert-manager/cert-manager-545d4d4674-6kmxt"] Mar 12 00:20:52 crc kubenswrapper[4948]: I0312 00:20:52.662572 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-545d4d4674-6kmxt" Mar 12 00:20:52 crc kubenswrapper[4948]: I0312 00:20:52.667516 4948 reflector.go:368] Caches populated for *v1.Secret from object-"cert-manager"/"cert-manager-dockercfg-2b2rk" Mar 12 00:20:52 crc kubenswrapper[4948]: I0312 00:20:52.715989 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-545d4d4674-6kmxt"] Mar 12 00:20:52 crc kubenswrapper[4948]: I0312 00:20:52.856143 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b61251ac-1f6d-4ddd-92c1-dc910d384274-bound-sa-token\") pod \"cert-manager-545d4d4674-6kmxt\" (UID: \"b61251ac-1f6d-4ddd-92c1-dc910d384274\") " pod="cert-manager/cert-manager-545d4d4674-6kmxt" Mar 12 00:20:52 crc kubenswrapper[4948]: I0312 00:20:52.856377 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-7p6sq\" (UniqueName: \"kubernetes.io/projected/b61251ac-1f6d-4ddd-92c1-dc910d384274-kube-api-access-7p6sq\") pod \"cert-manager-545d4d4674-6kmxt\" (UID: \"b61251ac-1f6d-4ddd-92c1-dc910d384274\") " pod="cert-manager/cert-manager-545d4d4674-6kmxt" Mar 12 00:20:52 crc kubenswrapper[4948]: I0312 00:20:52.957345 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-7p6sq\" (UniqueName: \"kubernetes.io/projected/b61251ac-1f6d-4ddd-92c1-dc910d384274-kube-api-access-7p6sq\") pod \"cert-manager-545d4d4674-6kmxt\" (UID: \"b61251ac-1f6d-4ddd-92c1-dc910d384274\") " pod="cert-manager/cert-manager-545d4d4674-6kmxt" Mar 12 00:20:52 crc kubenswrapper[4948]: I0312 00:20:52.957440 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b61251ac-1f6d-4ddd-92c1-dc910d384274-bound-sa-token\") pod \"cert-manager-545d4d4674-6kmxt\" (UID: \"b61251ac-1f6d-4ddd-92c1-dc910d384274\") " pod="cert-manager/cert-manager-545d4d4674-6kmxt" Mar 12 00:20:52 crc kubenswrapper[4948]: I0312 00:20:52.978887 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"bound-sa-token\" (UniqueName: \"kubernetes.io/projected/b61251ac-1f6d-4ddd-92c1-dc910d384274-bound-sa-token\") pod \"cert-manager-545d4d4674-6kmxt\" (UID: \"b61251ac-1f6d-4ddd-92c1-dc910d384274\") " pod="cert-manager/cert-manager-545d4d4674-6kmxt" Mar 12 00:20:52 crc kubenswrapper[4948]: I0312 00:20:52.992761 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-7p6sq\" (UniqueName: \"kubernetes.io/projected/b61251ac-1f6d-4ddd-92c1-dc910d384274-kube-api-access-7p6sq\") pod \"cert-manager-545d4d4674-6kmxt\" (UID: \"b61251ac-1f6d-4ddd-92c1-dc910d384274\") " pod="cert-manager/cert-manager-545d4d4674-6kmxt" Mar 12 00:20:53 crc kubenswrapper[4948]: I0312 00:20:53.285636 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="cert-manager/cert-manager-545d4d4674-6kmxt" Mar 12 00:20:53 crc kubenswrapper[4948]: I0312 00:20:53.795858 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["cert-manager/cert-manager-545d4d4674-6kmxt"] Mar 12 00:20:54 crc kubenswrapper[4948]: I0312 00:20:54.168180 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-545d4d4674-6kmxt" event={"ID":"b61251ac-1f6d-4ddd-92c1-dc910d384274","Type":"ContainerStarted","Data":"27d6788e4587bab7e3fb5188f88be365d8f662287ee6ae4ab9c1c8a23206a1b8"} Mar 12 00:20:54 crc kubenswrapper[4948]: I0312 00:20:54.168582 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="cert-manager/cert-manager-545d4d4674-6kmxt" event={"ID":"b61251ac-1f6d-4ddd-92c1-dc910d384274","Type":"ContainerStarted","Data":"3ccca7793a60cfeea142aa93e4bc291cd94db29b69fbeaa50d6bf4f03098d715"} Mar 12 00:20:54 crc kubenswrapper[4948]: I0312 00:20:54.634147 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="cert-manager/cert-manager-webhook-6888856db4-5qkhc" Mar 12 00:20:54 crc kubenswrapper[4948]: I0312 00:20:54.653823 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="cert-manager/cert-manager-545d4d4674-6kmxt" podStartSLOduration=2.6537975190000003 podStartE2EDuration="2.653797519s" podCreationTimestamp="2026-03-12 00:20:52 +0000 UTC" firstStartedPulling="0001-01-01 00:00:00 +0000 UTC" lastFinishedPulling="0001-01-01 00:00:00 +0000 UTC" observedRunningTime="2026-03-12 00:20:54.194358896 +0000 UTC m=+853.649962674" watchObservedRunningTime="2026-03-12 00:20:54.653797519 +0000 UTC m=+854.109401267" Mar 12 00:20:56 crc kubenswrapper[4948]: I0312 00:20:56.933438 4948 scope.go:117] "RemoveContainer" containerID="7a35745a45555a1ae1f5c1b4600afa7ae8e2ab7e7001b750cee96f3375eb7f42" Mar 12 00:21:00 crc kubenswrapper[4948]: I0312 00:21:00.341572 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="service-telemetry/elasticsearch-es-default-0" Mar 12 00:21:14 crc kubenswrapper[4948]: I0312 00:21:14.785127 4948 patch_prober.go:28] interesting pod/machine-config-daemon-m4xwc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 00:21:14 crc kubenswrapper[4948]: I0312 00:21:14.785744 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" podUID="7e483c2b-08f0-4e92-8e4a-b7281f30af3e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 00:21:44 crc kubenswrapper[4948]: I0312 00:21:44.784942 4948 patch_prober.go:28] interesting pod/machine-config-daemon-m4xwc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 00:21:44 crc kubenswrapper[4948]: I0312 00:21:44.785549 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" podUID="7e483c2b-08f0-4e92-8e4a-b7281f30af3e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 00:21:44 crc kubenswrapper[4948]: I0312 00:21:44.785600 4948 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" Mar 12 00:21:44 crc kubenswrapper[4948]: I0312 00:21:44.786222 4948 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"d57d486c11ea6dd0bb65ccd254f325841bcd048227a6896a9722dd323a6190db"} pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 12 00:21:44 crc kubenswrapper[4948]: I0312 00:21:44.786289 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" podUID="7e483c2b-08f0-4e92-8e4a-b7281f30af3e" containerName="machine-config-daemon" containerID="cri-o://d57d486c11ea6dd0bb65ccd254f325841bcd048227a6896a9722dd323a6190db" gracePeriod=600 Mar 12 00:21:45 crc kubenswrapper[4948]: I0312 00:21:45.540419 4948 generic.go:334] "Generic (PLEG): container finished" podID="7e483c2b-08f0-4e92-8e4a-b7281f30af3e" containerID="d57d486c11ea6dd0bb65ccd254f325841bcd048227a6896a9722dd323a6190db" exitCode=0 Mar 12 00:21:45 crc kubenswrapper[4948]: I0312 00:21:45.540896 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" event={"ID":"7e483c2b-08f0-4e92-8e4a-b7281f30af3e","Type":"ContainerDied","Data":"d57d486c11ea6dd0bb65ccd254f325841bcd048227a6896a9722dd323a6190db"} Mar 12 00:21:45 crc kubenswrapper[4948]: I0312 00:21:45.540926 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" event={"ID":"7e483c2b-08f0-4e92-8e4a-b7281f30af3e","Type":"ContainerStarted","Data":"8f5f07d17ca670886abddff8cc1ea5d57cf97a7dc2b3744fe1219d1f6e701f88"} Mar 12 00:21:45 crc kubenswrapper[4948]: I0312 00:21:45.540946 4948 scope.go:117] "RemoveContainer" containerID="4bae24b53bedc5b7f174f77eb51d6a9cf0630fe1120ad3885ae2181e65d6ec55" Mar 12 00:22:00 crc kubenswrapper[4948]: I0312 00:22:00.146736 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29554582-4h2q8"] Mar 12 00:22:00 crc kubenswrapper[4948]: I0312 00:22:00.147865 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29554582-4h2q8" Mar 12 00:22:00 crc kubenswrapper[4948]: I0312 00:22:00.151967 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-44g5m" Mar 12 00:22:00 crc kubenswrapper[4948]: I0312 00:22:00.152039 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 00:22:00 crc kubenswrapper[4948]: I0312 00:22:00.152097 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 00:22:00 crc kubenswrapper[4948]: I0312 00:22:00.157247 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29554582-4h2q8"] Mar 12 00:22:00 crc kubenswrapper[4948]: I0312 00:22:00.280776 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-4vgrw\" (UniqueName: \"kubernetes.io/projected/98e5bf0e-b218-4f0c-a23f-afdb5ef99139-kube-api-access-4vgrw\") pod \"auto-csr-approver-29554582-4h2q8\" (UID: \"98e5bf0e-b218-4f0c-a23f-afdb5ef99139\") " pod="openshift-infra/auto-csr-approver-29554582-4h2q8" Mar 12 00:22:00 crc kubenswrapper[4948]: I0312 00:22:00.382068 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-4vgrw\" (UniqueName: \"kubernetes.io/projected/98e5bf0e-b218-4f0c-a23f-afdb5ef99139-kube-api-access-4vgrw\") pod \"auto-csr-approver-29554582-4h2q8\" (UID: \"98e5bf0e-b218-4f0c-a23f-afdb5ef99139\") " pod="openshift-infra/auto-csr-approver-29554582-4h2q8" Mar 12 00:22:00 crc kubenswrapper[4948]: I0312 00:22:00.408878 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-4vgrw\" (UniqueName: \"kubernetes.io/projected/98e5bf0e-b218-4f0c-a23f-afdb5ef99139-kube-api-access-4vgrw\") pod \"auto-csr-approver-29554582-4h2q8\" (UID: \"98e5bf0e-b218-4f0c-a23f-afdb5ef99139\") " pod="openshift-infra/auto-csr-approver-29554582-4h2q8" Mar 12 00:22:00 crc kubenswrapper[4948]: I0312 00:22:00.476861 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29554582-4h2q8" Mar 12 00:22:00 crc kubenswrapper[4948]: I0312 00:22:00.735329 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29554582-4h2q8"] Mar 12 00:22:00 crc kubenswrapper[4948]: W0312 00:22:00.746029 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod98e5bf0e_b218_4f0c_a23f_afdb5ef99139.slice/crio-c86e0a4c67980c0c29570e9dca159b304359722c069e497b6d3f2f01efb0797f WatchSource:0}: Error finding container c86e0a4c67980c0c29570e9dca159b304359722c069e497b6d3f2f01efb0797f: Status 404 returned error can't find the container with id c86e0a4c67980c0c29570e9dca159b304359722c069e497b6d3f2f01efb0797f Mar 12 00:22:00 crc kubenswrapper[4948]: I0312 00:22:00.749363 4948 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 12 00:22:01 crc kubenswrapper[4948]: I0312 00:22:01.664664 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29554582-4h2q8" event={"ID":"98e5bf0e-b218-4f0c-a23f-afdb5ef99139","Type":"ContainerStarted","Data":"c86e0a4c67980c0c29570e9dca159b304359722c069e497b6d3f2f01efb0797f"} Mar 12 00:22:02 crc kubenswrapper[4948]: I0312 00:22:02.675008 4948 generic.go:334] "Generic (PLEG): container finished" podID="98e5bf0e-b218-4f0c-a23f-afdb5ef99139" containerID="78ae7ff1e7adb8335eb10046583835dd6e974d34218d47686d62115ff0197d49" exitCode=0 Mar 12 00:22:02 crc kubenswrapper[4948]: I0312 00:22:02.675107 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29554582-4h2q8" event={"ID":"98e5bf0e-b218-4f0c-a23f-afdb5ef99139","Type":"ContainerDied","Data":"78ae7ff1e7adb8335eb10046583835dd6e974d34218d47686d62115ff0197d49"} Mar 12 00:22:04 crc kubenswrapper[4948]: I0312 00:22:04.001166 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29554582-4h2q8" Mar 12 00:22:04 crc kubenswrapper[4948]: I0312 00:22:04.144382 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-4vgrw\" (UniqueName: \"kubernetes.io/projected/98e5bf0e-b218-4f0c-a23f-afdb5ef99139-kube-api-access-4vgrw\") pod \"98e5bf0e-b218-4f0c-a23f-afdb5ef99139\" (UID: \"98e5bf0e-b218-4f0c-a23f-afdb5ef99139\") " Mar 12 00:22:04 crc kubenswrapper[4948]: I0312 00:22:04.152907 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/98e5bf0e-b218-4f0c-a23f-afdb5ef99139-kube-api-access-4vgrw" (OuterVolumeSpecName: "kube-api-access-4vgrw") pod "98e5bf0e-b218-4f0c-a23f-afdb5ef99139" (UID: "98e5bf0e-b218-4f0c-a23f-afdb5ef99139"). InnerVolumeSpecName "kube-api-access-4vgrw". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:22:04 crc kubenswrapper[4948]: I0312 00:22:04.245756 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-4vgrw\" (UniqueName: \"kubernetes.io/projected/98e5bf0e-b218-4f0c-a23f-afdb5ef99139-kube-api-access-4vgrw\") on node \"crc\" DevicePath \"\"" Mar 12 00:22:04 crc kubenswrapper[4948]: I0312 00:22:04.692449 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29554582-4h2q8" event={"ID":"98e5bf0e-b218-4f0c-a23f-afdb5ef99139","Type":"ContainerDied","Data":"c86e0a4c67980c0c29570e9dca159b304359722c069e497b6d3f2f01efb0797f"} Mar 12 00:22:04 crc kubenswrapper[4948]: I0312 00:22:04.692493 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="c86e0a4c67980c0c29570e9dca159b304359722c069e497b6d3f2f01efb0797f" Mar 12 00:22:04 crc kubenswrapper[4948]: I0312 00:22:04.692548 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29554582-4h2q8" Mar 12 00:22:05 crc kubenswrapper[4948]: I0312 00:22:05.091647 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29554576-k9gx6"] Mar 12 00:22:05 crc kubenswrapper[4948]: I0312 00:22:05.097278 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29554576-k9gx6"] Mar 12 00:22:05 crc kubenswrapper[4948]: I0312 00:22:05.326299 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="551726c9-76b6-435a-ae73-c43821a95491" path="/var/lib/kubelet/pods/551726c9-76b6-435a-ae73-c43821a95491/volumes" Mar 12 00:22:24 crc kubenswrapper[4948]: I0312 00:22:24.720010 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-must-gather-gtx7f/must-gather-2r98w"] Mar 12 00:22:24 crc kubenswrapper[4948]: E0312 00:22:24.720725 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="98e5bf0e-b218-4f0c-a23f-afdb5ef99139" containerName="oc" Mar 12 00:22:24 crc kubenswrapper[4948]: I0312 00:22:24.720737 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="98e5bf0e-b218-4f0c-a23f-afdb5ef99139" containerName="oc" Mar 12 00:22:24 crc kubenswrapper[4948]: I0312 00:22:24.720850 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="98e5bf0e-b218-4f0c-a23f-afdb5ef99139" containerName="oc" Mar 12 00:22:24 crc kubenswrapper[4948]: I0312 00:22:24.721527 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gtx7f/must-gather-2r98w" Mar 12 00:22:24 crc kubenswrapper[4948]: I0312 00:22:24.723072 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-gtx7f"/"kube-root-ca.crt" Mar 12 00:22:24 crc kubenswrapper[4948]: I0312 00:22:24.723131 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-must-gather-gtx7f"/"openshift-service-ca.crt" Mar 12 00:22:24 crc kubenswrapper[4948]: I0312 00:22:24.729560 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-gtx7f/must-gather-2r98w"] Mar 12 00:22:24 crc kubenswrapper[4948]: I0312 00:22:24.844933 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/154d5896-889c-432a-b3b9-38f403f4c58c-must-gather-output\") pod \"must-gather-2r98w\" (UID: \"154d5896-889c-432a-b3b9-38f403f4c58c\") " pod="openshift-must-gather-gtx7f/must-gather-2r98w" Mar 12 00:22:24 crc kubenswrapper[4948]: I0312 00:22:24.844995 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-csknq\" (UniqueName: \"kubernetes.io/projected/154d5896-889c-432a-b3b9-38f403f4c58c-kube-api-access-csknq\") pod \"must-gather-2r98w\" (UID: \"154d5896-889c-432a-b3b9-38f403f4c58c\") " pod="openshift-must-gather-gtx7f/must-gather-2r98w" Mar 12 00:22:24 crc kubenswrapper[4948]: I0312 00:22:24.945955 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-csknq\" (UniqueName: \"kubernetes.io/projected/154d5896-889c-432a-b3b9-38f403f4c58c-kube-api-access-csknq\") pod \"must-gather-2r98w\" (UID: \"154d5896-889c-432a-b3b9-38f403f4c58c\") " pod="openshift-must-gather-gtx7f/must-gather-2r98w" Mar 12 00:22:24 crc kubenswrapper[4948]: I0312 00:22:24.946063 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/154d5896-889c-432a-b3b9-38f403f4c58c-must-gather-output\") pod \"must-gather-2r98w\" (UID: \"154d5896-889c-432a-b3b9-38f403f4c58c\") " pod="openshift-must-gather-gtx7f/must-gather-2r98w" Mar 12 00:22:24 crc kubenswrapper[4948]: I0312 00:22:24.946501 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/154d5896-889c-432a-b3b9-38f403f4c58c-must-gather-output\") pod \"must-gather-2r98w\" (UID: \"154d5896-889c-432a-b3b9-38f403f4c58c\") " pod="openshift-must-gather-gtx7f/must-gather-2r98w" Mar 12 00:22:24 crc kubenswrapper[4948]: I0312 00:22:24.968855 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-csknq\" (UniqueName: \"kubernetes.io/projected/154d5896-889c-432a-b3b9-38f403f4c58c-kube-api-access-csknq\") pod \"must-gather-2r98w\" (UID: \"154d5896-889c-432a-b3b9-38f403f4c58c\") " pod="openshift-must-gather-gtx7f/must-gather-2r98w" Mar 12 00:22:25 crc kubenswrapper[4948]: I0312 00:22:25.049630 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gtx7f/must-gather-2r98w" Mar 12 00:22:25 crc kubenswrapper[4948]: I0312 00:22:25.566072 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-must-gather-gtx7f/must-gather-2r98w"] Mar 12 00:22:25 crc kubenswrapper[4948]: I0312 00:22:25.852138 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gtx7f/must-gather-2r98w" event={"ID":"154d5896-889c-432a-b3b9-38f403f4c58c","Type":"ContainerStarted","Data":"9a95cf70d25c6ec46ee3a4a6867bb4e55a94ca5a5a4857886f54a87cb032b067"} Mar 12 00:22:32 crc kubenswrapper[4948]: I0312 00:22:32.929608 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gtx7f/must-gather-2r98w" event={"ID":"154d5896-889c-432a-b3b9-38f403f4c58c","Type":"ContainerStarted","Data":"a917e6d6943c7a3ff5ad629931488f955662bfc9af179535531abf117f1402bb"} Mar 12 00:22:32 crc kubenswrapper[4948]: I0312 00:22:32.930132 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gtx7f/must-gather-2r98w" event={"ID":"154d5896-889c-432a-b3b9-38f403f4c58c","Type":"ContainerStarted","Data":"6b76568a00c880e6ad8c9df0426826ff23ca81056481883223fb9e66c3033747"} Mar 12 00:22:32 crc kubenswrapper[4948]: I0312 00:22:32.956912 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-must-gather-gtx7f/must-gather-2r98w" podStartSLOduration=2.812433064 podStartE2EDuration="8.956894486s" podCreationTimestamp="2026-03-12 00:22:24 +0000 UTC" firstStartedPulling="2026-03-12 00:22:25.577882311 +0000 UTC m=+945.033486049" lastFinishedPulling="2026-03-12 00:22:31.722343733 +0000 UTC m=+951.177947471" observedRunningTime="2026-03-12 00:22:32.956668751 +0000 UTC m=+952.412272489" watchObservedRunningTime="2026-03-12 00:22:32.956894486 +0000 UTC m=+952.412498225" Mar 12 00:22:57 crc kubenswrapper[4948]: I0312 00:22:57.023754 4948 scope.go:117] "RemoveContainer" containerID="cd53f999f5fd5a2b763323f1d841e6ceed15b820af87bb68737992dda04f42c2" Mar 12 00:23:03 crc kubenswrapper[4948]: I0312 00:23:03.014070 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/community-operators-fct2p"] Mar 12 00:23:03 crc kubenswrapper[4948]: I0312 00:23:03.017252 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fct2p" Mar 12 00:23:03 crc kubenswrapper[4948]: I0312 00:23:03.032077 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fct2p"] Mar 12 00:23:03 crc kubenswrapper[4948]: I0312 00:23:03.061359 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2de2883d-63eb-4a04-9af9-9f2139b0e7f0-utilities\") pod \"community-operators-fct2p\" (UID: \"2de2883d-63eb-4a04-9af9-9f2139b0e7f0\") " pod="openshift-marketplace/community-operators-fct2p" Mar 12 00:23:03 crc kubenswrapper[4948]: I0312 00:23:03.061422 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2de2883d-63eb-4a04-9af9-9f2139b0e7f0-catalog-content\") pod \"community-operators-fct2p\" (UID: \"2de2883d-63eb-4a04-9af9-9f2139b0e7f0\") " pod="openshift-marketplace/community-operators-fct2p" Mar 12 00:23:03 crc kubenswrapper[4948]: I0312 00:23:03.061493 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-cq5hv\" (UniqueName: \"kubernetes.io/projected/2de2883d-63eb-4a04-9af9-9f2139b0e7f0-kube-api-access-cq5hv\") pod \"community-operators-fct2p\" (UID: \"2de2883d-63eb-4a04-9af9-9f2139b0e7f0\") " pod="openshift-marketplace/community-operators-fct2p" Mar 12 00:23:03 crc kubenswrapper[4948]: I0312 00:23:03.162321 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2de2883d-63eb-4a04-9af9-9f2139b0e7f0-utilities\") pod \"community-operators-fct2p\" (UID: \"2de2883d-63eb-4a04-9af9-9f2139b0e7f0\") " pod="openshift-marketplace/community-operators-fct2p" Mar 12 00:23:03 crc kubenswrapper[4948]: I0312 00:23:03.162383 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2de2883d-63eb-4a04-9af9-9f2139b0e7f0-catalog-content\") pod \"community-operators-fct2p\" (UID: \"2de2883d-63eb-4a04-9af9-9f2139b0e7f0\") " pod="openshift-marketplace/community-operators-fct2p" Mar 12 00:23:03 crc kubenswrapper[4948]: I0312 00:23:03.162451 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-cq5hv\" (UniqueName: \"kubernetes.io/projected/2de2883d-63eb-4a04-9af9-9f2139b0e7f0-kube-api-access-cq5hv\") pod \"community-operators-fct2p\" (UID: \"2de2883d-63eb-4a04-9af9-9f2139b0e7f0\") " pod="openshift-marketplace/community-operators-fct2p" Mar 12 00:23:03 crc kubenswrapper[4948]: I0312 00:23:03.162979 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2de2883d-63eb-4a04-9af9-9f2139b0e7f0-utilities\") pod \"community-operators-fct2p\" (UID: \"2de2883d-63eb-4a04-9af9-9f2139b0e7f0\") " pod="openshift-marketplace/community-operators-fct2p" Mar 12 00:23:03 crc kubenswrapper[4948]: I0312 00:23:03.163053 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2de2883d-63eb-4a04-9af9-9f2139b0e7f0-catalog-content\") pod \"community-operators-fct2p\" (UID: \"2de2883d-63eb-4a04-9af9-9f2139b0e7f0\") " pod="openshift-marketplace/community-operators-fct2p" Mar 12 00:23:03 crc kubenswrapper[4948]: I0312 00:23:03.190164 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-cq5hv\" (UniqueName: \"kubernetes.io/projected/2de2883d-63eb-4a04-9af9-9f2139b0e7f0-kube-api-access-cq5hv\") pod \"community-operators-fct2p\" (UID: \"2de2883d-63eb-4a04-9af9-9f2139b0e7f0\") " pod="openshift-marketplace/community-operators-fct2p" Mar 12 00:23:03 crc kubenswrapper[4948]: I0312 00:23:03.347029 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fct2p" Mar 12 00:23:03 crc kubenswrapper[4948]: I0312 00:23:03.794105 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/community-operators-fct2p"] Mar 12 00:23:04 crc kubenswrapper[4948]: I0312 00:23:04.175273 4948 generic.go:334] "Generic (PLEG): container finished" podID="2de2883d-63eb-4a04-9af9-9f2139b0e7f0" containerID="a8150090d68de8dd1b2f94c30e882d2b4b4e3099318793dbd92cdea1ba40929c" exitCode=0 Mar 12 00:23:04 crc kubenswrapper[4948]: I0312 00:23:04.175351 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fct2p" event={"ID":"2de2883d-63eb-4a04-9af9-9f2139b0e7f0","Type":"ContainerDied","Data":"a8150090d68de8dd1b2f94c30e882d2b4b4e3099318793dbd92cdea1ba40929c"} Mar 12 00:23:04 crc kubenswrapper[4948]: I0312 00:23:04.176192 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fct2p" event={"ID":"2de2883d-63eb-4a04-9af9-9f2139b0e7f0","Type":"ContainerStarted","Data":"785d0315fe660170535fa904f292dc4fd7ebb7b77bb06ed0607b9a5f4c1ca048"} Mar 12 00:23:06 crc kubenswrapper[4948]: I0312 00:23:06.193254 4948 generic.go:334] "Generic (PLEG): container finished" podID="2de2883d-63eb-4a04-9af9-9f2139b0e7f0" containerID="fa4f96bb77f4b94c5033783df18bbe2767c93c8fc1a48a157af73650330c79cb" exitCode=0 Mar 12 00:23:06 crc kubenswrapper[4948]: I0312 00:23:06.193484 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fct2p" event={"ID":"2de2883d-63eb-4a04-9af9-9f2139b0e7f0","Type":"ContainerDied","Data":"fa4f96bb77f4b94c5033783df18bbe2767c93c8fc1a48a157af73650330c79cb"} Mar 12 00:23:07 crc kubenswrapper[4948]: I0312 00:23:07.201728 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fct2p" event={"ID":"2de2883d-63eb-4a04-9af9-9f2139b0e7f0","Type":"ContainerStarted","Data":"cb316d6253b8b8a3677bbec2a248cd56d3d156d711d806159728a589ca377762"} Mar 12 00:23:07 crc kubenswrapper[4948]: I0312 00:23:07.220156 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/community-operators-fct2p" podStartSLOduration=2.749405502 podStartE2EDuration="5.220135239s" podCreationTimestamp="2026-03-12 00:23:02 +0000 UTC" firstStartedPulling="2026-03-12 00:23:04.177469728 +0000 UTC m=+983.633073456" lastFinishedPulling="2026-03-12 00:23:06.648199445 +0000 UTC m=+986.103803193" observedRunningTime="2026-03-12 00:23:07.217349462 +0000 UTC m=+986.672953200" watchObservedRunningTime="2026-03-12 00:23:07.220135239 +0000 UTC m=+986.675738997" Mar 12 00:23:11 crc kubenswrapper[4948]: I0312 00:23:11.844576 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/redhat-operators-stpqf"] Mar 12 00:23:11 crc kubenswrapper[4948]: I0312 00:23:11.847032 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-stpqf" Mar 12 00:23:11 crc kubenswrapper[4948]: I0312 00:23:11.853223 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-stpqf"] Mar 12 00:23:11 crc kubenswrapper[4948]: I0312 00:23:11.883667 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd3ab5d7-6789-4fc8-b838-6a18f3d5683e-utilities\") pod \"redhat-operators-stpqf\" (UID: \"bd3ab5d7-6789-4fc8-b838-6a18f3d5683e\") " pod="openshift-marketplace/redhat-operators-stpqf" Mar 12 00:23:11 crc kubenswrapper[4948]: I0312 00:23:11.883796 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd3ab5d7-6789-4fc8-b838-6a18f3d5683e-catalog-content\") pod \"redhat-operators-stpqf\" (UID: \"bd3ab5d7-6789-4fc8-b838-6a18f3d5683e\") " pod="openshift-marketplace/redhat-operators-stpqf" Mar 12 00:23:11 crc kubenswrapper[4948]: I0312 00:23:11.883828 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-882c8\" (UniqueName: \"kubernetes.io/projected/bd3ab5d7-6789-4fc8-b838-6a18f3d5683e-kube-api-access-882c8\") pod \"redhat-operators-stpqf\" (UID: \"bd3ab5d7-6789-4fc8-b838-6a18f3d5683e\") " pod="openshift-marketplace/redhat-operators-stpqf" Mar 12 00:23:11 crc kubenswrapper[4948]: I0312 00:23:11.984992 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd3ab5d7-6789-4fc8-b838-6a18f3d5683e-catalog-content\") pod \"redhat-operators-stpqf\" (UID: \"bd3ab5d7-6789-4fc8-b838-6a18f3d5683e\") " pod="openshift-marketplace/redhat-operators-stpqf" Mar 12 00:23:11 crc kubenswrapper[4948]: I0312 00:23:11.985063 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-882c8\" (UniqueName: \"kubernetes.io/projected/bd3ab5d7-6789-4fc8-b838-6a18f3d5683e-kube-api-access-882c8\") pod \"redhat-operators-stpqf\" (UID: \"bd3ab5d7-6789-4fc8-b838-6a18f3d5683e\") " pod="openshift-marketplace/redhat-operators-stpqf" Mar 12 00:23:11 crc kubenswrapper[4948]: I0312 00:23:11.985128 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd3ab5d7-6789-4fc8-b838-6a18f3d5683e-utilities\") pod \"redhat-operators-stpqf\" (UID: \"bd3ab5d7-6789-4fc8-b838-6a18f3d5683e\") " pod="openshift-marketplace/redhat-operators-stpqf" Mar 12 00:23:11 crc kubenswrapper[4948]: I0312 00:23:11.985529 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd3ab5d7-6789-4fc8-b838-6a18f3d5683e-catalog-content\") pod \"redhat-operators-stpqf\" (UID: \"bd3ab5d7-6789-4fc8-b838-6a18f3d5683e\") " pod="openshift-marketplace/redhat-operators-stpqf" Mar 12 00:23:11 crc kubenswrapper[4948]: I0312 00:23:11.985633 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd3ab5d7-6789-4fc8-b838-6a18f3d5683e-utilities\") pod \"redhat-operators-stpqf\" (UID: \"bd3ab5d7-6789-4fc8-b838-6a18f3d5683e\") " pod="openshift-marketplace/redhat-operators-stpqf" Mar 12 00:23:12 crc kubenswrapper[4948]: I0312 00:23:12.007971 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-882c8\" (UniqueName: \"kubernetes.io/projected/bd3ab5d7-6789-4fc8-b838-6a18f3d5683e-kube-api-access-882c8\") pod \"redhat-operators-stpqf\" (UID: \"bd3ab5d7-6789-4fc8-b838-6a18f3d5683e\") " pod="openshift-marketplace/redhat-operators-stpqf" Mar 12 00:23:12 crc kubenswrapper[4948]: I0312 00:23:12.181665 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-stpqf" Mar 12 00:23:12 crc kubenswrapper[4948]: I0312 00:23:12.587685 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/redhat-operators-stpqf"] Mar 12 00:23:13 crc kubenswrapper[4948]: I0312 00:23:13.239422 4948 generic.go:334] "Generic (PLEG): container finished" podID="bd3ab5d7-6789-4fc8-b838-6a18f3d5683e" containerID="b08f67fc1052016732f76c986bee93f13cdb852b472a6ffd019ec44fb91ef7e8" exitCode=0 Mar 12 00:23:13 crc kubenswrapper[4948]: I0312 00:23:13.239515 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-stpqf" event={"ID":"bd3ab5d7-6789-4fc8-b838-6a18f3d5683e","Type":"ContainerDied","Data":"b08f67fc1052016732f76c986bee93f13cdb852b472a6ffd019ec44fb91ef7e8"} Mar 12 00:23:13 crc kubenswrapper[4948]: I0312 00:23:13.239659 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-stpqf" event={"ID":"bd3ab5d7-6789-4fc8-b838-6a18f3d5683e","Type":"ContainerStarted","Data":"49fdee0aa5ed3e9efa8928f71f2d901ddf45c00a83fdf39d9809aa5d5d8262b0"} Mar 12 00:23:13 crc kubenswrapper[4948]: I0312 00:23:13.347372 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/community-operators-fct2p" Mar 12 00:23:13 crc kubenswrapper[4948]: I0312 00:23:13.347475 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/community-operators-fct2p" Mar 12 00:23:13 crc kubenswrapper[4948]: I0312 00:23:13.397606 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/community-operators-fct2p" Mar 12 00:23:14 crc kubenswrapper[4948]: I0312 00:23:14.252197 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-stpqf" event={"ID":"bd3ab5d7-6789-4fc8-b838-6a18f3d5683e","Type":"ContainerStarted","Data":"4cb14541c900d7fde977fddfa43627bcf447f008fe9d399d5ac89b364c976a7a"} Mar 12 00:23:14 crc kubenswrapper[4948]: I0312 00:23:14.299692 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/community-operators-fct2p" Mar 12 00:23:15 crc kubenswrapper[4948]: I0312 00:23:15.257578 4948 generic.go:334] "Generic (PLEG): container finished" podID="bd3ab5d7-6789-4fc8-b838-6a18f3d5683e" containerID="4cb14541c900d7fde977fddfa43627bcf447f008fe9d399d5ac89b364c976a7a" exitCode=0 Mar 12 00:23:15 crc kubenswrapper[4948]: I0312 00:23:15.257642 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-stpqf" event={"ID":"bd3ab5d7-6789-4fc8-b838-6a18f3d5683e","Type":"ContainerDied","Data":"4cb14541c900d7fde977fddfa43627bcf447f008fe9d399d5ac89b364c976a7a"} Mar 12 00:23:15 crc kubenswrapper[4948]: I0312 00:23:15.812563 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fct2p"] Mar 12 00:23:16 crc kubenswrapper[4948]: I0312 00:23:16.264125 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-stpqf" event={"ID":"bd3ab5d7-6789-4fc8-b838-6a18f3d5683e","Type":"ContainerStarted","Data":"69b4679f77f44c8f5b45b15864c777fe167911e652b85d2fc0928554e50d30db"} Mar 12 00:23:16 crc kubenswrapper[4948]: I0312 00:23:16.281151 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/redhat-operators-stpqf" podStartSLOduration=2.849912365 podStartE2EDuration="5.281131505s" podCreationTimestamp="2026-03-12 00:23:11 +0000 UTC" firstStartedPulling="2026-03-12 00:23:13.240912802 +0000 UTC m=+992.696516540" lastFinishedPulling="2026-03-12 00:23:15.672131902 +0000 UTC m=+995.127735680" observedRunningTime="2026-03-12 00:23:16.27912162 +0000 UTC m=+995.734725358" watchObservedRunningTime="2026-03-12 00:23:16.281131505 +0000 UTC m=+995.736735243" Mar 12 00:23:17 crc kubenswrapper[4948]: I0312 00:23:17.269419 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/community-operators-fct2p" podUID="2de2883d-63eb-4a04-9af9-9f2139b0e7f0" containerName="registry-server" containerID="cri-o://cb316d6253b8b8a3677bbec2a248cd56d3d156d711d806159728a589ca377762" gracePeriod=2 Mar 12 00:23:17 crc kubenswrapper[4948]: I0312 00:23:17.653273 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fct2p" Mar 12 00:23:17 crc kubenswrapper[4948]: I0312 00:23:17.757544 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-cq5hv\" (UniqueName: \"kubernetes.io/projected/2de2883d-63eb-4a04-9af9-9f2139b0e7f0-kube-api-access-cq5hv\") pod \"2de2883d-63eb-4a04-9af9-9f2139b0e7f0\" (UID: \"2de2883d-63eb-4a04-9af9-9f2139b0e7f0\") " Mar 12 00:23:17 crc kubenswrapper[4948]: I0312 00:23:17.757656 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2de2883d-63eb-4a04-9af9-9f2139b0e7f0-utilities\") pod \"2de2883d-63eb-4a04-9af9-9f2139b0e7f0\" (UID: \"2de2883d-63eb-4a04-9af9-9f2139b0e7f0\") " Mar 12 00:23:17 crc kubenswrapper[4948]: I0312 00:23:17.757733 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2de2883d-63eb-4a04-9af9-9f2139b0e7f0-catalog-content\") pod \"2de2883d-63eb-4a04-9af9-9f2139b0e7f0\" (UID: \"2de2883d-63eb-4a04-9af9-9f2139b0e7f0\") " Mar 12 00:23:17 crc kubenswrapper[4948]: I0312 00:23:17.761065 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2de2883d-63eb-4a04-9af9-9f2139b0e7f0-utilities" (OuterVolumeSpecName: "utilities") pod "2de2883d-63eb-4a04-9af9-9f2139b0e7f0" (UID: "2de2883d-63eb-4a04-9af9-9f2139b0e7f0"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 00:23:17 crc kubenswrapper[4948]: I0312 00:23:17.772461 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/2de2883d-63eb-4a04-9af9-9f2139b0e7f0-kube-api-access-cq5hv" (OuterVolumeSpecName: "kube-api-access-cq5hv") pod "2de2883d-63eb-4a04-9af9-9f2139b0e7f0" (UID: "2de2883d-63eb-4a04-9af9-9f2139b0e7f0"). InnerVolumeSpecName "kube-api-access-cq5hv". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:23:17 crc kubenswrapper[4948]: I0312 00:23:17.805635 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/2de2883d-63eb-4a04-9af9-9f2139b0e7f0-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "2de2883d-63eb-4a04-9af9-9f2139b0e7f0" (UID: "2de2883d-63eb-4a04-9af9-9f2139b0e7f0"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 00:23:17 crc kubenswrapper[4948]: I0312 00:23:17.859134 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/2de2883d-63eb-4a04-9af9-9f2139b0e7f0-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 00:23:17 crc kubenswrapper[4948]: I0312 00:23:17.859380 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/2de2883d-63eb-4a04-9af9-9f2139b0e7f0-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 00:23:17 crc kubenswrapper[4948]: I0312 00:23:17.859451 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-cq5hv\" (UniqueName: \"kubernetes.io/projected/2de2883d-63eb-4a04-9af9-9f2139b0e7f0-kube-api-access-cq5hv\") on node \"crc\" DevicePath \"\"" Mar 12 00:23:18 crc kubenswrapper[4948]: I0312 00:23:18.278469 4948 generic.go:334] "Generic (PLEG): container finished" podID="2de2883d-63eb-4a04-9af9-9f2139b0e7f0" containerID="cb316d6253b8b8a3677bbec2a248cd56d3d156d711d806159728a589ca377762" exitCode=0 Mar 12 00:23:18 crc kubenswrapper[4948]: I0312 00:23:18.278520 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fct2p" event={"ID":"2de2883d-63eb-4a04-9af9-9f2139b0e7f0","Type":"ContainerDied","Data":"cb316d6253b8b8a3677bbec2a248cd56d3d156d711d806159728a589ca377762"} Mar 12 00:23:18 crc kubenswrapper[4948]: I0312 00:23:18.279255 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/community-operators-fct2p" event={"ID":"2de2883d-63eb-4a04-9af9-9f2139b0e7f0","Type":"ContainerDied","Data":"785d0315fe660170535fa904f292dc4fd7ebb7b77bb06ed0607b9a5f4c1ca048"} Mar 12 00:23:18 crc kubenswrapper[4948]: I0312 00:23:18.278550 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/community-operators-fct2p" Mar 12 00:23:18 crc kubenswrapper[4948]: I0312 00:23:18.279345 4948 scope.go:117] "RemoveContainer" containerID="cb316d6253b8b8a3677bbec2a248cd56d3d156d711d806159728a589ca377762" Mar 12 00:23:18 crc kubenswrapper[4948]: I0312 00:23:18.307553 4948 scope.go:117] "RemoveContainer" containerID="fa4f96bb77f4b94c5033783df18bbe2767c93c8fc1a48a157af73650330c79cb" Mar 12 00:23:18 crc kubenswrapper[4948]: I0312 00:23:18.333596 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/community-operators-fct2p"] Mar 12 00:23:18 crc kubenswrapper[4948]: I0312 00:23:18.337992 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/community-operators-fct2p"] Mar 12 00:23:18 crc kubenswrapper[4948]: I0312 00:23:18.347450 4948 scope.go:117] "RemoveContainer" containerID="a8150090d68de8dd1b2f94c30e882d2b4b4e3099318793dbd92cdea1ba40929c" Mar 12 00:23:18 crc kubenswrapper[4948]: I0312 00:23:18.366127 4948 scope.go:117] "RemoveContainer" containerID="cb316d6253b8b8a3677bbec2a248cd56d3d156d711d806159728a589ca377762" Mar 12 00:23:18 crc kubenswrapper[4948]: E0312 00:23:18.366786 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"cb316d6253b8b8a3677bbec2a248cd56d3d156d711d806159728a589ca377762\": container with ID starting with cb316d6253b8b8a3677bbec2a248cd56d3d156d711d806159728a589ca377762 not found: ID does not exist" containerID="cb316d6253b8b8a3677bbec2a248cd56d3d156d711d806159728a589ca377762" Mar 12 00:23:18 crc kubenswrapper[4948]: I0312 00:23:18.366847 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"cb316d6253b8b8a3677bbec2a248cd56d3d156d711d806159728a589ca377762"} err="failed to get container status \"cb316d6253b8b8a3677bbec2a248cd56d3d156d711d806159728a589ca377762\": rpc error: code = NotFound desc = could not find container \"cb316d6253b8b8a3677bbec2a248cd56d3d156d711d806159728a589ca377762\": container with ID starting with cb316d6253b8b8a3677bbec2a248cd56d3d156d711d806159728a589ca377762 not found: ID does not exist" Mar 12 00:23:18 crc kubenswrapper[4948]: I0312 00:23:18.366884 4948 scope.go:117] "RemoveContainer" containerID="fa4f96bb77f4b94c5033783df18bbe2767c93c8fc1a48a157af73650330c79cb" Mar 12 00:23:18 crc kubenswrapper[4948]: E0312 00:23:18.367232 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"fa4f96bb77f4b94c5033783df18bbe2767c93c8fc1a48a157af73650330c79cb\": container with ID starting with fa4f96bb77f4b94c5033783df18bbe2767c93c8fc1a48a157af73650330c79cb not found: ID does not exist" containerID="fa4f96bb77f4b94c5033783df18bbe2767c93c8fc1a48a157af73650330c79cb" Mar 12 00:23:18 crc kubenswrapper[4948]: I0312 00:23:18.367274 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"fa4f96bb77f4b94c5033783df18bbe2767c93c8fc1a48a157af73650330c79cb"} err="failed to get container status \"fa4f96bb77f4b94c5033783df18bbe2767c93c8fc1a48a157af73650330c79cb\": rpc error: code = NotFound desc = could not find container \"fa4f96bb77f4b94c5033783df18bbe2767c93c8fc1a48a157af73650330c79cb\": container with ID starting with fa4f96bb77f4b94c5033783df18bbe2767c93c8fc1a48a157af73650330c79cb not found: ID does not exist" Mar 12 00:23:18 crc kubenswrapper[4948]: I0312 00:23:18.367342 4948 scope.go:117] "RemoveContainer" containerID="a8150090d68de8dd1b2f94c30e882d2b4b4e3099318793dbd92cdea1ba40929c" Mar 12 00:23:18 crc kubenswrapper[4948]: E0312 00:23:18.367643 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"a8150090d68de8dd1b2f94c30e882d2b4b4e3099318793dbd92cdea1ba40929c\": container with ID starting with a8150090d68de8dd1b2f94c30e882d2b4b4e3099318793dbd92cdea1ba40929c not found: ID does not exist" containerID="a8150090d68de8dd1b2f94c30e882d2b4b4e3099318793dbd92cdea1ba40929c" Mar 12 00:23:18 crc kubenswrapper[4948]: I0312 00:23:18.367679 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"a8150090d68de8dd1b2f94c30e882d2b4b4e3099318793dbd92cdea1ba40929c"} err="failed to get container status \"a8150090d68de8dd1b2f94c30e882d2b4b4e3099318793dbd92cdea1ba40929c\": rpc error: code = NotFound desc = could not find container \"a8150090d68de8dd1b2f94c30e882d2b4b4e3099318793dbd92cdea1ba40929c\": container with ID starting with a8150090d68de8dd1b2f94c30e882d2b4b4e3099318793dbd92cdea1ba40929c not found: ID does not exist" Mar 12 00:23:19 crc kubenswrapper[4948]: I0312 00:23:19.324578 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="2de2883d-63eb-4a04-9af9-9f2139b0e7f0" path="/var/lib/kubelet/pods/2de2883d-63eb-4a04-9af9-9f2139b0e7f0/volumes" Mar 12 00:23:20 crc kubenswrapper[4948]: I0312 00:23:20.337080 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_control-plane-machine-set-operator-78cbb6b69f-bj72v_2352fbc0-94ff-4b0c-b499-d16b94c279df/control-plane-machine-set-operator/0.log" Mar 12 00:23:20 crc kubenswrapper[4948]: I0312 00:23:20.491366 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-g7g94_a544acdb-b758-4ea8-a43d-8270fc454901/machine-api-operator/0.log" Mar 12 00:23:20 crc kubenswrapper[4948]: I0312 00:23:20.516758 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-machine-api_machine-api-operator-5694c8668f-g7g94_a544acdb-b758-4ea8-a43d-8270fc454901/kube-rbac-proxy/0.log" Mar 12 00:23:22 crc kubenswrapper[4948]: I0312 00:23:22.182367 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/redhat-operators-stpqf" Mar 12 00:23:22 crc kubenswrapper[4948]: I0312 00:23:22.182717 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/redhat-operators-stpqf" Mar 12 00:23:23 crc kubenswrapper[4948]: I0312 00:23:23.228872 4948 prober.go:107] "Probe failed" probeType="Startup" pod="openshift-marketplace/redhat-operators-stpqf" podUID="bd3ab5d7-6789-4fc8-b838-6a18f3d5683e" containerName="registry-server" probeResult="failure" output=< Mar 12 00:23:23 crc kubenswrapper[4948]: timeout: failed to connect service ":50051" within 1s Mar 12 00:23:23 crc kubenswrapper[4948]: > Mar 12 00:23:32 crc kubenswrapper[4948]: I0312 00:23:32.252718 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/redhat-operators-stpqf" Mar 12 00:23:32 crc kubenswrapper[4948]: I0312 00:23:32.322226 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/redhat-operators-stpqf" Mar 12 00:23:32 crc kubenswrapper[4948]: I0312 00:23:32.490886 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-stpqf"] Mar 12 00:23:33 crc kubenswrapper[4948]: I0312 00:23:33.386976 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/redhat-operators-stpqf" podUID="bd3ab5d7-6789-4fc8-b838-6a18f3d5683e" containerName="registry-server" containerID="cri-o://69b4679f77f44c8f5b45b15864c777fe167911e652b85d2fc0928554e50d30db" gracePeriod=2 Mar 12 00:23:33 crc kubenswrapper[4948]: I0312 00:23:33.737047 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-stpqf" Mar 12 00:23:33 crc kubenswrapper[4948]: I0312 00:23:33.777352 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-545d4d4674-6kmxt_b61251ac-1f6d-4ddd-92c1-dc910d384274/cert-manager-controller/0.log" Mar 12 00:23:33 crc kubenswrapper[4948]: I0312 00:23:33.883597 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd3ab5d7-6789-4fc8-b838-6a18f3d5683e-utilities\") pod \"bd3ab5d7-6789-4fc8-b838-6a18f3d5683e\" (UID: \"bd3ab5d7-6789-4fc8-b838-6a18f3d5683e\") " Mar 12 00:23:33 crc kubenswrapper[4948]: I0312 00:23:33.883691 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd3ab5d7-6789-4fc8-b838-6a18f3d5683e-catalog-content\") pod \"bd3ab5d7-6789-4fc8-b838-6a18f3d5683e\" (UID: \"bd3ab5d7-6789-4fc8-b838-6a18f3d5683e\") " Mar 12 00:23:33 crc kubenswrapper[4948]: I0312 00:23:33.883740 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-882c8\" (UniqueName: \"kubernetes.io/projected/bd3ab5d7-6789-4fc8-b838-6a18f3d5683e-kube-api-access-882c8\") pod \"bd3ab5d7-6789-4fc8-b838-6a18f3d5683e\" (UID: \"bd3ab5d7-6789-4fc8-b838-6a18f3d5683e\") " Mar 12 00:23:33 crc kubenswrapper[4948]: I0312 00:23:33.884563 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd3ab5d7-6789-4fc8-b838-6a18f3d5683e-utilities" (OuterVolumeSpecName: "utilities") pod "bd3ab5d7-6789-4fc8-b838-6a18f3d5683e" (UID: "bd3ab5d7-6789-4fc8-b838-6a18f3d5683e"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 00:23:33 crc kubenswrapper[4948]: I0312 00:23:33.890967 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/bd3ab5d7-6789-4fc8-b838-6a18f3d5683e-kube-api-access-882c8" (OuterVolumeSpecName: "kube-api-access-882c8") pod "bd3ab5d7-6789-4fc8-b838-6a18f3d5683e" (UID: "bd3ab5d7-6789-4fc8-b838-6a18f3d5683e"). InnerVolumeSpecName "kube-api-access-882c8". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:23:33 crc kubenswrapper[4948]: I0312 00:23:33.967859 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-cainjector-5545bd876-7bjwb_dcc3f47a-6e31-417d-ab68-f4a2d50b4556/cert-manager-cainjector/0.log" Mar 12 00:23:33 crc kubenswrapper[4948]: I0312 00:23:33.983967 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/cert-manager_cert-manager-webhook-6888856db4-5qkhc_50d01335-9cbd-4637-9657-27936e2265b1/cert-manager-webhook/0.log" Mar 12 00:23:33 crc kubenswrapper[4948]: I0312 00:23:33.985133 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/bd3ab5d7-6789-4fc8-b838-6a18f3d5683e-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 00:23:33 crc kubenswrapper[4948]: I0312 00:23:33.985159 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-882c8\" (UniqueName: \"kubernetes.io/projected/bd3ab5d7-6789-4fc8-b838-6a18f3d5683e-kube-api-access-882c8\") on node \"crc\" DevicePath \"\"" Mar 12 00:23:34 crc kubenswrapper[4948]: I0312 00:23:34.007880 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/bd3ab5d7-6789-4fc8-b838-6a18f3d5683e-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "bd3ab5d7-6789-4fc8-b838-6a18f3d5683e" (UID: "bd3ab5d7-6789-4fc8-b838-6a18f3d5683e"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 00:23:34 crc kubenswrapper[4948]: I0312 00:23:34.086276 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/bd3ab5d7-6789-4fc8-b838-6a18f3d5683e-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 00:23:34 crc kubenswrapper[4948]: I0312 00:23:34.394203 4948 generic.go:334] "Generic (PLEG): container finished" podID="bd3ab5d7-6789-4fc8-b838-6a18f3d5683e" containerID="69b4679f77f44c8f5b45b15864c777fe167911e652b85d2fc0928554e50d30db" exitCode=0 Mar 12 00:23:34 crc kubenswrapper[4948]: I0312 00:23:34.394240 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-stpqf" event={"ID":"bd3ab5d7-6789-4fc8-b838-6a18f3d5683e","Type":"ContainerDied","Data":"69b4679f77f44c8f5b45b15864c777fe167911e652b85d2fc0928554e50d30db"} Mar 12 00:23:34 crc kubenswrapper[4948]: I0312 00:23:34.394265 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/redhat-operators-stpqf" event={"ID":"bd3ab5d7-6789-4fc8-b838-6a18f3d5683e","Type":"ContainerDied","Data":"49fdee0aa5ed3e9efa8928f71f2d901ddf45c00a83fdf39d9809aa5d5d8262b0"} Mar 12 00:23:34 crc kubenswrapper[4948]: I0312 00:23:34.394281 4948 scope.go:117] "RemoveContainer" containerID="69b4679f77f44c8f5b45b15864c777fe167911e652b85d2fc0928554e50d30db" Mar 12 00:23:34 crc kubenswrapper[4948]: I0312 00:23:34.394419 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/redhat-operators-stpqf" Mar 12 00:23:34 crc kubenswrapper[4948]: I0312 00:23:34.424257 4948 scope.go:117] "RemoveContainer" containerID="4cb14541c900d7fde977fddfa43627bcf447f008fe9d399d5ac89b364c976a7a" Mar 12 00:23:34 crc kubenswrapper[4948]: I0312 00:23:34.444128 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/redhat-operators-stpqf"] Mar 12 00:23:34 crc kubenswrapper[4948]: I0312 00:23:34.446549 4948 scope.go:117] "RemoveContainer" containerID="b08f67fc1052016732f76c986bee93f13cdb852b472a6ffd019ec44fb91ef7e8" Mar 12 00:23:34 crc kubenswrapper[4948]: I0312 00:23:34.448555 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/redhat-operators-stpqf"] Mar 12 00:23:34 crc kubenswrapper[4948]: I0312 00:23:34.473646 4948 scope.go:117] "RemoveContainer" containerID="69b4679f77f44c8f5b45b15864c777fe167911e652b85d2fc0928554e50d30db" Mar 12 00:23:34 crc kubenswrapper[4948]: E0312 00:23:34.474170 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"69b4679f77f44c8f5b45b15864c777fe167911e652b85d2fc0928554e50d30db\": container with ID starting with 69b4679f77f44c8f5b45b15864c777fe167911e652b85d2fc0928554e50d30db not found: ID does not exist" containerID="69b4679f77f44c8f5b45b15864c777fe167911e652b85d2fc0928554e50d30db" Mar 12 00:23:34 crc kubenswrapper[4948]: I0312 00:23:34.474228 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"69b4679f77f44c8f5b45b15864c777fe167911e652b85d2fc0928554e50d30db"} err="failed to get container status \"69b4679f77f44c8f5b45b15864c777fe167911e652b85d2fc0928554e50d30db\": rpc error: code = NotFound desc = could not find container \"69b4679f77f44c8f5b45b15864c777fe167911e652b85d2fc0928554e50d30db\": container with ID starting with 69b4679f77f44c8f5b45b15864c777fe167911e652b85d2fc0928554e50d30db not found: ID does not exist" Mar 12 00:23:34 crc kubenswrapper[4948]: I0312 00:23:34.474252 4948 scope.go:117] "RemoveContainer" containerID="4cb14541c900d7fde977fddfa43627bcf447f008fe9d399d5ac89b364c976a7a" Mar 12 00:23:34 crc kubenswrapper[4948]: E0312 00:23:34.474728 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"4cb14541c900d7fde977fddfa43627bcf447f008fe9d399d5ac89b364c976a7a\": container with ID starting with 4cb14541c900d7fde977fddfa43627bcf447f008fe9d399d5ac89b364c976a7a not found: ID does not exist" containerID="4cb14541c900d7fde977fddfa43627bcf447f008fe9d399d5ac89b364c976a7a" Mar 12 00:23:34 crc kubenswrapper[4948]: I0312 00:23:34.474816 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"4cb14541c900d7fde977fddfa43627bcf447f008fe9d399d5ac89b364c976a7a"} err="failed to get container status \"4cb14541c900d7fde977fddfa43627bcf447f008fe9d399d5ac89b364c976a7a\": rpc error: code = NotFound desc = could not find container \"4cb14541c900d7fde977fddfa43627bcf447f008fe9d399d5ac89b364c976a7a\": container with ID starting with 4cb14541c900d7fde977fddfa43627bcf447f008fe9d399d5ac89b364c976a7a not found: ID does not exist" Mar 12 00:23:34 crc kubenswrapper[4948]: I0312 00:23:34.474881 4948 scope.go:117] "RemoveContainer" containerID="b08f67fc1052016732f76c986bee93f13cdb852b472a6ffd019ec44fb91ef7e8" Mar 12 00:23:34 crc kubenswrapper[4948]: E0312 00:23:34.475446 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b08f67fc1052016732f76c986bee93f13cdb852b472a6ffd019ec44fb91ef7e8\": container with ID starting with b08f67fc1052016732f76c986bee93f13cdb852b472a6ffd019ec44fb91ef7e8 not found: ID does not exist" containerID="b08f67fc1052016732f76c986bee93f13cdb852b472a6ffd019ec44fb91ef7e8" Mar 12 00:23:34 crc kubenswrapper[4948]: I0312 00:23:34.475507 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b08f67fc1052016732f76c986bee93f13cdb852b472a6ffd019ec44fb91ef7e8"} err="failed to get container status \"b08f67fc1052016732f76c986bee93f13cdb852b472a6ffd019ec44fb91ef7e8\": rpc error: code = NotFound desc = could not find container \"b08f67fc1052016732f76c986bee93f13cdb852b472a6ffd019ec44fb91ef7e8\": container with ID starting with b08f67fc1052016732f76c986bee93f13cdb852b472a6ffd019ec44fb91ef7e8 not found: ID does not exist" Mar 12 00:23:35 crc kubenswrapper[4948]: I0312 00:23:35.325584 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="bd3ab5d7-6789-4fc8-b838-6a18f3d5683e" path="/var/lib/kubelet/pods/bd3ab5d7-6789-4fc8-b838-6a18f3d5683e/volumes" Mar 12 00:23:48 crc kubenswrapper[4948]: I0312 00:23:48.929592 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-68bc856cb9-cvw7f_b5f03616-ebd8-4698-91f7-ca2419301e08/prometheus-operator/0.log" Mar 12 00:23:49 crc kubenswrapper[4948]: I0312 00:23:49.071212 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-85fd558b5-d55ck_06125956-526b-468b-b609-9bb13c834499/prometheus-operator-admission-webhook/0.log" Mar 12 00:23:49 crc kubenswrapper[4948]: I0312 00:23:49.097913 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-85fd558b5-jvl2f_406df0e7-9930-440b-9023-5369cdac443b/prometheus-operator-admission-webhook/0.log" Mar 12 00:23:49 crc kubenswrapper[4948]: I0312 00:23:49.230260 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-59bdc8b94-bkzp4_bf64ec59-b44a-407b-8d10-f1f9bd91de76/operator/0.log" Mar 12 00:23:49 crc kubenswrapper[4948]: I0312 00:23:49.278390 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5bf474d74f-swd7r_f7223c5e-0d07-4513-a05a-3d504647724c/perses-operator/0.log" Mar 12 00:24:00 crc kubenswrapper[4948]: I0312 00:24:00.151535 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29554584-5mxvf"] Mar 12 00:24:00 crc kubenswrapper[4948]: E0312 00:24:00.152356 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd3ab5d7-6789-4fc8-b838-6a18f3d5683e" containerName="registry-server" Mar 12 00:24:00 crc kubenswrapper[4948]: I0312 00:24:00.152373 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd3ab5d7-6789-4fc8-b838-6a18f3d5683e" containerName="registry-server" Mar 12 00:24:00 crc kubenswrapper[4948]: E0312 00:24:00.152394 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2de2883d-63eb-4a04-9af9-9f2139b0e7f0" containerName="extract-utilities" Mar 12 00:24:00 crc kubenswrapper[4948]: I0312 00:24:00.152402 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="2de2883d-63eb-4a04-9af9-9f2139b0e7f0" containerName="extract-utilities" Mar 12 00:24:00 crc kubenswrapper[4948]: E0312 00:24:00.152418 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2de2883d-63eb-4a04-9af9-9f2139b0e7f0" containerName="extract-content" Mar 12 00:24:00 crc kubenswrapper[4948]: I0312 00:24:00.152427 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="2de2883d-63eb-4a04-9af9-9f2139b0e7f0" containerName="extract-content" Mar 12 00:24:00 crc kubenswrapper[4948]: E0312 00:24:00.152438 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd3ab5d7-6789-4fc8-b838-6a18f3d5683e" containerName="extract-utilities" Mar 12 00:24:00 crc kubenswrapper[4948]: I0312 00:24:00.152445 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd3ab5d7-6789-4fc8-b838-6a18f3d5683e" containerName="extract-utilities" Mar 12 00:24:00 crc kubenswrapper[4948]: E0312 00:24:00.152461 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="2de2883d-63eb-4a04-9af9-9f2139b0e7f0" containerName="registry-server" Mar 12 00:24:00 crc kubenswrapper[4948]: I0312 00:24:00.152469 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="2de2883d-63eb-4a04-9af9-9f2139b0e7f0" containerName="registry-server" Mar 12 00:24:00 crc kubenswrapper[4948]: E0312 00:24:00.152481 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="bd3ab5d7-6789-4fc8-b838-6a18f3d5683e" containerName="extract-content" Mar 12 00:24:00 crc kubenswrapper[4948]: I0312 00:24:00.152489 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="bd3ab5d7-6789-4fc8-b838-6a18f3d5683e" containerName="extract-content" Mar 12 00:24:00 crc kubenswrapper[4948]: I0312 00:24:00.152609 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="bd3ab5d7-6789-4fc8-b838-6a18f3d5683e" containerName="registry-server" Mar 12 00:24:00 crc kubenswrapper[4948]: I0312 00:24:00.152622 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="2de2883d-63eb-4a04-9af9-9f2139b0e7f0" containerName="registry-server" Mar 12 00:24:00 crc kubenswrapper[4948]: I0312 00:24:00.153087 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29554584-5mxvf" Mar 12 00:24:00 crc kubenswrapper[4948]: I0312 00:24:00.156118 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 00:24:00 crc kubenswrapper[4948]: I0312 00:24:00.156453 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-44g5m" Mar 12 00:24:00 crc kubenswrapper[4948]: I0312 00:24:00.164363 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 00:24:00 crc kubenswrapper[4948]: I0312 00:24:00.165852 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29554584-5mxvf"] Mar 12 00:24:00 crc kubenswrapper[4948]: I0312 00:24:00.169206 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-jblgc\" (UniqueName: \"kubernetes.io/projected/8689cab2-961f-4987-b988-00b097f80961-kube-api-access-jblgc\") pod \"auto-csr-approver-29554584-5mxvf\" (UID: \"8689cab2-961f-4987-b988-00b097f80961\") " pod="openshift-infra/auto-csr-approver-29554584-5mxvf" Mar 12 00:24:00 crc kubenswrapper[4948]: I0312 00:24:00.270668 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-jblgc\" (UniqueName: \"kubernetes.io/projected/8689cab2-961f-4987-b988-00b097f80961-kube-api-access-jblgc\") pod \"auto-csr-approver-29554584-5mxvf\" (UID: \"8689cab2-961f-4987-b988-00b097f80961\") " pod="openshift-infra/auto-csr-approver-29554584-5mxvf" Mar 12 00:24:00 crc kubenswrapper[4948]: I0312 00:24:00.302674 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-jblgc\" (UniqueName: \"kubernetes.io/projected/8689cab2-961f-4987-b988-00b097f80961-kube-api-access-jblgc\") pod \"auto-csr-approver-29554584-5mxvf\" (UID: \"8689cab2-961f-4987-b988-00b097f80961\") " pod="openshift-infra/auto-csr-approver-29554584-5mxvf" Mar 12 00:24:00 crc kubenswrapper[4948]: I0312 00:24:00.485406 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29554584-5mxvf" Mar 12 00:24:01 crc kubenswrapper[4948]: I0312 00:24:01.010962 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29554584-5mxvf"] Mar 12 00:24:01 crc kubenswrapper[4948]: I0312 00:24:01.606191 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29554584-5mxvf" event={"ID":"8689cab2-961f-4987-b988-00b097f80961","Type":"ContainerStarted","Data":"5434cb1f5614933fdcaab1912f47669dcd4b7a5df2e3ec78cac0a571f06f4b7e"} Mar 12 00:24:03 crc kubenswrapper[4948]: I0312 00:24:03.626421 4948 generic.go:334] "Generic (PLEG): container finished" podID="8689cab2-961f-4987-b988-00b097f80961" containerID="f490ecb0c2b62324b87c623c1d51240d63ed35762aea1a751b21e15d56849249" exitCode=0 Mar 12 00:24:03 crc kubenswrapper[4948]: I0312 00:24:03.626503 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29554584-5mxvf" event={"ID":"8689cab2-961f-4987-b988-00b097f80961","Type":"ContainerDied","Data":"f490ecb0c2b62324b87c623c1d51240d63ed35762aea1a751b21e15d56849249"} Mar 12 00:24:04 crc kubenswrapper[4948]: I0312 00:24:04.229135 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn6w4j_d1dc348c-ee53-42f5-aebb-e482dff1ba79/util/0.log" Mar 12 00:24:04 crc kubenswrapper[4948]: I0312 00:24:04.387940 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn6w4j_d1dc348c-ee53-42f5-aebb-e482dff1ba79/util/0.log" Mar 12 00:24:04 crc kubenswrapper[4948]: I0312 00:24:04.402692 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn6w4j_d1dc348c-ee53-42f5-aebb-e482dff1ba79/pull/0.log" Mar 12 00:24:04 crc kubenswrapper[4948]: I0312 00:24:04.439892 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn6w4j_d1dc348c-ee53-42f5-aebb-e482dff1ba79/pull/0.log" Mar 12 00:24:04 crc kubenswrapper[4948]: I0312 00:24:04.583202 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn6w4j_d1dc348c-ee53-42f5-aebb-e482dff1ba79/pull/0.log" Mar 12 00:24:04 crc kubenswrapper[4948]: I0312 00:24:04.593614 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn6w4j_d1dc348c-ee53-42f5-aebb-e482dff1ba79/util/0.log" Mar 12 00:24:04 crc kubenswrapper[4948]: I0312 00:24:04.595775 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_6e3e74c24700cc2bb66271d960117ff0976dc779e6a3bc37905b952e8fn6w4j_d1dc348c-ee53-42f5-aebb-e482dff1ba79/extract/0.log" Mar 12 00:24:04 crc kubenswrapper[4948]: I0312 00:24:04.770736 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39e9cm9g_539fe10f-66ae-4110-a85e-0f42febbf096/util/0.log" Mar 12 00:24:04 crc kubenswrapper[4948]: I0312 00:24:04.879983 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29554584-5mxvf" Mar 12 00:24:04 crc kubenswrapper[4948]: I0312 00:24:04.929399 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-jblgc\" (UniqueName: \"kubernetes.io/projected/8689cab2-961f-4987-b988-00b097f80961-kube-api-access-jblgc\") pod \"8689cab2-961f-4987-b988-00b097f80961\" (UID: \"8689cab2-961f-4987-b988-00b097f80961\") " Mar 12 00:24:04 crc kubenswrapper[4948]: I0312 00:24:04.935597 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/8689cab2-961f-4987-b988-00b097f80961-kube-api-access-jblgc" (OuterVolumeSpecName: "kube-api-access-jblgc") pod "8689cab2-961f-4987-b988-00b097f80961" (UID: "8689cab2-961f-4987-b988-00b097f80961"). InnerVolumeSpecName "kube-api-access-jblgc". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:24:04 crc kubenswrapper[4948]: I0312 00:24:04.955824 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39e9cm9g_539fe10f-66ae-4110-a85e-0f42febbf096/util/0.log" Mar 12 00:24:04 crc kubenswrapper[4948]: I0312 00:24:04.957780 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39e9cm9g_539fe10f-66ae-4110-a85e-0f42febbf096/pull/0.log" Mar 12 00:24:04 crc kubenswrapper[4948]: I0312 00:24:04.972461 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39e9cm9g_539fe10f-66ae-4110-a85e-0f42febbf096/pull/0.log" Mar 12 00:24:05 crc kubenswrapper[4948]: I0312 00:24:05.030907 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-jblgc\" (UniqueName: \"kubernetes.io/projected/8689cab2-961f-4987-b988-00b097f80961-kube-api-access-jblgc\") on node \"crc\" DevicePath \"\"" Mar 12 00:24:05 crc kubenswrapper[4948]: I0312 00:24:05.106123 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39e9cm9g_539fe10f-66ae-4110-a85e-0f42febbf096/util/0.log" Mar 12 00:24:05 crc kubenswrapper[4948]: I0312 00:24:05.132211 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39e9cm9g_539fe10f-66ae-4110-a85e-0f42febbf096/pull/0.log" Mar 12 00:24:05 crc kubenswrapper[4948]: I0312 00:24:05.184589 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_7acef1e4a10e04db4e216682ff91f6a23804f55f83b8dd8f8f8f5ac39e9cm9g_539fe10f-66ae-4110-a85e-0f42febbf096/extract/0.log" Mar 12 00:24:05 crc kubenswrapper[4948]: I0312 00:24:05.273981 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5j9jsp_3aa75b6b-35be-412c-8fa5-1e16b61a40fc/util/0.log" Mar 12 00:24:05 crc kubenswrapper[4948]: I0312 00:24:05.401336 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5j9jsp_3aa75b6b-35be-412c-8fa5-1e16b61a40fc/pull/0.log" Mar 12 00:24:05 crc kubenswrapper[4948]: I0312 00:24:05.432093 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5j9jsp_3aa75b6b-35be-412c-8fa5-1e16b61a40fc/util/0.log" Mar 12 00:24:05 crc kubenswrapper[4948]: I0312 00:24:05.457900 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5j9jsp_3aa75b6b-35be-412c-8fa5-1e16b61a40fc/pull/0.log" Mar 12 00:24:05 crc kubenswrapper[4948]: I0312 00:24:05.607186 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5j9jsp_3aa75b6b-35be-412c-8fa5-1e16b61a40fc/util/0.log" Mar 12 00:24:05 crc kubenswrapper[4948]: I0312 00:24:05.623957 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5j9jsp_3aa75b6b-35be-412c-8fa5-1e16b61a40fc/extract/0.log" Mar 12 00:24:05 crc kubenswrapper[4948]: I0312 00:24:05.636817 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_925ad1f05bf386dc21bdfe2f8249c1fbfd04a404dec7a7fb6362d758e5j9jsp_3aa75b6b-35be-412c-8fa5-1e16b61a40fc/pull/0.log" Mar 12 00:24:05 crc kubenswrapper[4948]: I0312 00:24:05.639336 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29554584-5mxvf" event={"ID":"8689cab2-961f-4987-b988-00b097f80961","Type":"ContainerDied","Data":"5434cb1f5614933fdcaab1912f47669dcd4b7a5df2e3ec78cac0a571f06f4b7e"} Mar 12 00:24:05 crc kubenswrapper[4948]: I0312 00:24:05.639418 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29554584-5mxvf" Mar 12 00:24:05 crc kubenswrapper[4948]: I0312 00:24:05.639934 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="5434cb1f5614933fdcaab1912f47669dcd4b7a5df2e3ec78cac0a571f06f4b7e" Mar 12 00:24:05 crc kubenswrapper[4948]: I0312 00:24:05.765405 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc676_a2b1eac3-5989-4e06-b457-16991daaac81/util/0.log" Mar 12 00:24:05 crc kubenswrapper[4948]: I0312 00:24:05.931284 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29554578-fz9lc"] Mar 12 00:24:05 crc kubenswrapper[4948]: I0312 00:24:05.939572 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29554578-fz9lc"] Mar 12 00:24:05 crc kubenswrapper[4948]: I0312 00:24:05.940136 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc676_a2b1eac3-5989-4e06-b457-16991daaac81/util/0.log" Mar 12 00:24:05 crc kubenswrapper[4948]: I0312 00:24:05.940709 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc676_a2b1eac3-5989-4e06-b457-16991daaac81/pull/0.log" Mar 12 00:24:05 crc kubenswrapper[4948]: I0312 00:24:05.971181 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc676_a2b1eac3-5989-4e06-b457-16991daaac81/pull/0.log" Mar 12 00:24:06 crc kubenswrapper[4948]: I0312 00:24:06.126100 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc676_a2b1eac3-5989-4e06-b457-16991daaac81/pull/0.log" Mar 12 00:24:06 crc kubenswrapper[4948]: I0312 00:24:06.147135 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc676_a2b1eac3-5989-4e06-b457-16991daaac81/util/0.log" Mar 12 00:24:06 crc kubenswrapper[4948]: I0312 00:24:06.150222 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_98629960b44b381d1a86cff1d1439a8df43509c9ad24579158c59d0f08pc676_a2b1eac3-5989-4e06-b457-16991daaac81/extract/0.log" Mar 12 00:24:06 crc kubenswrapper[4948]: I0312 00:24:06.308068 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-zpktj_3fc280b5-e8c9-4046-bd99-74a0fcbc73ff/extract-utilities/0.log" Mar 12 00:24:06 crc kubenswrapper[4948]: I0312 00:24:06.461706 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-zpktj_3fc280b5-e8c9-4046-bd99-74a0fcbc73ff/extract-utilities/0.log" Mar 12 00:24:06 crc kubenswrapper[4948]: I0312 00:24:06.465621 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-zpktj_3fc280b5-e8c9-4046-bd99-74a0fcbc73ff/extract-content/0.log" Mar 12 00:24:06 crc kubenswrapper[4948]: I0312 00:24:06.466329 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-zpktj_3fc280b5-e8c9-4046-bd99-74a0fcbc73ff/extract-content/0.log" Mar 12 00:24:06 crc kubenswrapper[4948]: I0312 00:24:06.635723 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-zpktj_3fc280b5-e8c9-4046-bd99-74a0fcbc73ff/extract-content/0.log" Mar 12 00:24:06 crc kubenswrapper[4948]: I0312 00:24:06.641779 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-zpktj_3fc280b5-e8c9-4046-bd99-74a0fcbc73ff/extract-utilities/0.log" Mar 12 00:24:06 crc kubenswrapper[4948]: I0312 00:24:06.777408 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_certified-operators-zpktj_3fc280b5-e8c9-4046-bd99-74a0fcbc73ff/registry-server/0.log" Mar 12 00:24:06 crc kubenswrapper[4948]: I0312 00:24:06.811962 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dc79v_6658e419-c016-4b00-bb2c-24ffdbc65283/extract-utilities/0.log" Mar 12 00:24:07 crc kubenswrapper[4948]: I0312 00:24:07.018442 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dc79v_6658e419-c016-4b00-bb2c-24ffdbc65283/extract-utilities/0.log" Mar 12 00:24:07 crc kubenswrapper[4948]: I0312 00:24:07.028713 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dc79v_6658e419-c016-4b00-bb2c-24ffdbc65283/extract-content/0.log" Mar 12 00:24:07 crc kubenswrapper[4948]: I0312 00:24:07.040074 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dc79v_6658e419-c016-4b00-bb2c-24ffdbc65283/extract-content/0.log" Mar 12 00:24:07 crc kubenswrapper[4948]: I0312 00:24:07.247310 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dc79v_6658e419-c016-4b00-bb2c-24ffdbc65283/extract-utilities/0.log" Mar 12 00:24:07 crc kubenswrapper[4948]: I0312 00:24:07.257231 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dc79v_6658e419-c016-4b00-bb2c-24ffdbc65283/extract-content/0.log" Mar 12 00:24:07 crc kubenswrapper[4948]: I0312 00:24:07.324295 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="3b3f134e-8447-45bb-9b8e-ab8980e7a614" path="/var/lib/kubelet/pods/3b3f134e-8447-45bb-9b8e-ab8980e7a614/volumes" Mar 12 00:24:07 crc kubenswrapper[4948]: I0312 00:24:07.441096 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_marketplace-operator-79b997595-5k5ph_d39a6a83-8fa5-47e4-8047-4b4e713e1a64/marketplace-operator/0.log" Mar 12 00:24:07 crc kubenswrapper[4948]: I0312 00:24:07.460262 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_community-operators-dc79v_6658e419-c016-4b00-bb2c-24ffdbc65283/registry-server/0.log" Mar 12 00:24:07 crc kubenswrapper[4948]: I0312 00:24:07.536023 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vmmhs_da5b5036-fd96-4ac1-9567-46b309cdf00d/extract-utilities/0.log" Mar 12 00:24:07 crc kubenswrapper[4948]: I0312 00:24:07.635180 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vmmhs_da5b5036-fd96-4ac1-9567-46b309cdf00d/extract-content/0.log" Mar 12 00:24:07 crc kubenswrapper[4948]: I0312 00:24:07.662738 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vmmhs_da5b5036-fd96-4ac1-9567-46b309cdf00d/extract-utilities/0.log" Mar 12 00:24:07 crc kubenswrapper[4948]: I0312 00:24:07.691136 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vmmhs_da5b5036-fd96-4ac1-9567-46b309cdf00d/extract-content/0.log" Mar 12 00:24:07 crc kubenswrapper[4948]: I0312 00:24:07.839050 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vmmhs_da5b5036-fd96-4ac1-9567-46b309cdf00d/extract-content/0.log" Mar 12 00:24:07 crc kubenswrapper[4948]: I0312 00:24:07.841369 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vmmhs_da5b5036-fd96-4ac1-9567-46b309cdf00d/extract-utilities/0.log" Mar 12 00:24:07 crc kubenswrapper[4948]: I0312 00:24:07.952049 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-marketplace_redhat-operators-vmmhs_da5b5036-fd96-4ac1-9567-46b309cdf00d/registry-server/0.log" Mar 12 00:24:14 crc kubenswrapper[4948]: I0312 00:24:14.784702 4948 patch_prober.go:28] interesting pod/machine-config-daemon-m4xwc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 00:24:14 crc kubenswrapper[4948]: I0312 00:24:14.785081 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" podUID="7e483c2b-08f0-4e92-8e4a-b7281f30af3e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 00:24:21 crc kubenswrapper[4948]: I0312 00:24:21.594252 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-68bc856cb9-cvw7f_b5f03616-ebd8-4698-91f7-ca2419301e08/prometheus-operator/0.log" Mar 12 00:24:21 crc kubenswrapper[4948]: I0312 00:24:21.620467 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-85fd558b5-d55ck_06125956-526b-468b-b609-9bb13c834499/prometheus-operator-admission-webhook/0.log" Mar 12 00:24:21 crc kubenswrapper[4948]: I0312 00:24:21.645749 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_obo-prometheus-operator-admission-webhook-85fd558b5-jvl2f_406df0e7-9930-440b-9023-5369cdac443b/prometheus-operator-admission-webhook/0.log" Mar 12 00:24:21 crc kubenswrapper[4948]: I0312 00:24:21.706205 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_observability-operator-59bdc8b94-bkzp4_bf64ec59-b44a-407b-8d10-f1f9bd91de76/operator/0.log" Mar 12 00:24:21 crc kubenswrapper[4948]: I0312 00:24:21.881825 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-operators_perses-operator-5bf474d74f-swd7r_f7223c5e-0d07-4513-a05a-3d504647724c/perses-operator/0.log" Mar 12 00:24:44 crc kubenswrapper[4948]: I0312 00:24:44.785151 4948 patch_prober.go:28] interesting pod/machine-config-daemon-m4xwc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 00:24:44 crc kubenswrapper[4948]: I0312 00:24:44.786036 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" podUID="7e483c2b-08f0-4e92-8e4a-b7281f30af3e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 00:24:57 crc kubenswrapper[4948]: I0312 00:24:57.121633 4948 scope.go:117] "RemoveContainer" containerID="c4d3d32ce81f8386f0b5bf4319cfb711a12429b17af996353ce7b07372b881af" Mar 12 00:25:10 crc kubenswrapper[4948]: I0312 00:25:10.160774 4948 generic.go:334] "Generic (PLEG): container finished" podID="154d5896-889c-432a-b3b9-38f403f4c58c" containerID="6b76568a00c880e6ad8c9df0426826ff23ca81056481883223fb9e66c3033747" exitCode=0 Mar 12 00:25:10 crc kubenswrapper[4948]: I0312 00:25:10.160865 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-must-gather-gtx7f/must-gather-2r98w" event={"ID":"154d5896-889c-432a-b3b9-38f403f4c58c","Type":"ContainerDied","Data":"6b76568a00c880e6ad8c9df0426826ff23ca81056481883223fb9e66c3033747"} Mar 12 00:25:10 crc kubenswrapper[4948]: I0312 00:25:10.162029 4948 scope.go:117] "RemoveContainer" containerID="6b76568a00c880e6ad8c9df0426826ff23ca81056481883223fb9e66c3033747" Mar 12 00:25:11 crc kubenswrapper[4948]: I0312 00:25:11.165601 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-gtx7f_must-gather-2r98w_154d5896-889c-432a-b3b9-38f403f4c58c/gather/0.log" Mar 12 00:25:14 crc kubenswrapper[4948]: I0312 00:25:14.784894 4948 patch_prober.go:28] interesting pod/machine-config-daemon-m4xwc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 00:25:14 crc kubenswrapper[4948]: I0312 00:25:14.785595 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" podUID="7e483c2b-08f0-4e92-8e4a-b7281f30af3e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 00:25:14 crc kubenswrapper[4948]: I0312 00:25:14.785668 4948 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" Mar 12 00:25:14 crc kubenswrapper[4948]: I0312 00:25:14.786389 4948 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"8f5f07d17ca670886abddff8cc1ea5d57cf97a7dc2b3744fe1219d1f6e701f88"} pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 12 00:25:14 crc kubenswrapper[4948]: I0312 00:25:14.786487 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" podUID="7e483c2b-08f0-4e92-8e4a-b7281f30af3e" containerName="machine-config-daemon" containerID="cri-o://8f5f07d17ca670886abddff8cc1ea5d57cf97a7dc2b3744fe1219d1f6e701f88" gracePeriod=600 Mar 12 00:25:15 crc kubenswrapper[4948]: I0312 00:25:15.205351 4948 generic.go:334] "Generic (PLEG): container finished" podID="7e483c2b-08f0-4e92-8e4a-b7281f30af3e" containerID="8f5f07d17ca670886abddff8cc1ea5d57cf97a7dc2b3744fe1219d1f6e701f88" exitCode=0 Mar 12 00:25:15 crc kubenswrapper[4948]: I0312 00:25:15.205401 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" event={"ID":"7e483c2b-08f0-4e92-8e4a-b7281f30af3e","Type":"ContainerDied","Data":"8f5f07d17ca670886abddff8cc1ea5d57cf97a7dc2b3744fe1219d1f6e701f88"} Mar 12 00:25:15 crc kubenswrapper[4948]: I0312 00:25:15.205752 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" event={"ID":"7e483c2b-08f0-4e92-8e4a-b7281f30af3e","Type":"ContainerStarted","Data":"4207332ab030bc6bdba00683e8604b734b37d7a5dcb4073592c80c1f073959b8"} Mar 12 00:25:15 crc kubenswrapper[4948]: I0312 00:25:15.205788 4948 scope.go:117] "RemoveContainer" containerID="d57d486c11ea6dd0bb65ccd254f325841bcd048227a6896a9722dd323a6190db" Mar 12 00:25:18 crc kubenswrapper[4948]: I0312 00:25:18.065247 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-must-gather-gtx7f/must-gather-2r98w"] Mar 12 00:25:18 crc kubenswrapper[4948]: I0312 00:25:18.066374 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-must-gather-gtx7f/must-gather-2r98w" podUID="154d5896-889c-432a-b3b9-38f403f4c58c" containerName="copy" containerID="cri-o://a917e6d6943c7a3ff5ad629931488f955662bfc9af179535531abf117f1402bb" gracePeriod=2 Mar 12 00:25:18 crc kubenswrapper[4948]: I0312 00:25:18.079235 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-must-gather-gtx7f/must-gather-2r98w"] Mar 12 00:25:18 crc kubenswrapper[4948]: I0312 00:25:18.239172 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-gtx7f_must-gather-2r98w_154d5896-889c-432a-b3b9-38f403f4c58c/copy/0.log" Mar 12 00:25:18 crc kubenswrapper[4948]: I0312 00:25:18.240179 4948 generic.go:334] "Generic (PLEG): container finished" podID="154d5896-889c-432a-b3b9-38f403f4c58c" containerID="a917e6d6943c7a3ff5ad629931488f955662bfc9af179535531abf117f1402bb" exitCode=143 Mar 12 00:25:18 crc kubenswrapper[4948]: I0312 00:25:18.462263 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-gtx7f_must-gather-2r98w_154d5896-889c-432a-b3b9-38f403f4c58c/copy/0.log" Mar 12 00:25:18 crc kubenswrapper[4948]: I0312 00:25:18.463009 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gtx7f/must-gather-2r98w" Mar 12 00:25:18 crc kubenswrapper[4948]: I0312 00:25:18.526362 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/154d5896-889c-432a-b3b9-38f403f4c58c-must-gather-output\") pod \"154d5896-889c-432a-b3b9-38f403f4c58c\" (UID: \"154d5896-889c-432a-b3b9-38f403f4c58c\") " Mar 12 00:25:18 crc kubenswrapper[4948]: I0312 00:25:18.526427 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-csknq\" (UniqueName: \"kubernetes.io/projected/154d5896-889c-432a-b3b9-38f403f4c58c-kube-api-access-csknq\") pod \"154d5896-889c-432a-b3b9-38f403f4c58c\" (UID: \"154d5896-889c-432a-b3b9-38f403f4c58c\") " Mar 12 00:25:18 crc kubenswrapper[4948]: I0312 00:25:18.547066 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/154d5896-889c-432a-b3b9-38f403f4c58c-kube-api-access-csknq" (OuterVolumeSpecName: "kube-api-access-csknq") pod "154d5896-889c-432a-b3b9-38f403f4c58c" (UID: "154d5896-889c-432a-b3b9-38f403f4c58c"). InnerVolumeSpecName "kube-api-access-csknq". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:25:18 crc kubenswrapper[4948]: I0312 00:25:18.593499 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/154d5896-889c-432a-b3b9-38f403f4c58c-must-gather-output" (OuterVolumeSpecName: "must-gather-output") pod "154d5896-889c-432a-b3b9-38f403f4c58c" (UID: "154d5896-889c-432a-b3b9-38f403f4c58c"). InnerVolumeSpecName "must-gather-output". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 00:25:18 crc kubenswrapper[4948]: I0312 00:25:18.627900 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-csknq\" (UniqueName: \"kubernetes.io/projected/154d5896-889c-432a-b3b9-38f403f4c58c-kube-api-access-csknq\") on node \"crc\" DevicePath \"\"" Mar 12 00:25:18 crc kubenswrapper[4948]: I0312 00:25:18.627938 4948 reconciler_common.go:293] "Volume detached for volume \"must-gather-output\" (UniqueName: \"kubernetes.io/empty-dir/154d5896-889c-432a-b3b9-38f403f4c58c-must-gather-output\") on node \"crc\" DevicePath \"\"" Mar 12 00:25:19 crc kubenswrapper[4948]: I0312 00:25:19.251288 4948 log.go:25] "Finished parsing log file" path="/var/log/pods/openshift-must-gather-gtx7f_must-gather-2r98w_154d5896-889c-432a-b3b9-38f403f4c58c/copy/0.log" Mar 12 00:25:19 crc kubenswrapper[4948]: I0312 00:25:19.252222 4948 scope.go:117] "RemoveContainer" containerID="a917e6d6943c7a3ff5ad629931488f955662bfc9af179535531abf117f1402bb" Mar 12 00:25:19 crc kubenswrapper[4948]: I0312 00:25:19.252344 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-must-gather-gtx7f/must-gather-2r98w" Mar 12 00:25:19 crc kubenswrapper[4948]: I0312 00:25:19.285624 4948 scope.go:117] "RemoveContainer" containerID="6b76568a00c880e6ad8c9df0426826ff23ca81056481883223fb9e66c3033747" Mar 12 00:25:19 crc kubenswrapper[4948]: I0312 00:25:19.323670 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="154d5896-889c-432a-b3b9-38f403f4c58c" path="/var/lib/kubelet/pods/154d5896-889c-432a-b3b9-38f403f4c58c/volumes" Mar 12 00:25:57 crc kubenswrapper[4948]: I0312 00:25:57.271213 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-marketplace/certified-operators-8g2vb"] Mar 12 00:25:57 crc kubenswrapper[4948]: E0312 00:25:57.273573 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="154d5896-889c-432a-b3b9-38f403f4c58c" containerName="copy" Mar 12 00:25:57 crc kubenswrapper[4948]: I0312 00:25:57.273763 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="154d5896-889c-432a-b3b9-38f403f4c58c" containerName="copy" Mar 12 00:25:57 crc kubenswrapper[4948]: E0312 00:25:57.273788 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="8689cab2-961f-4987-b988-00b097f80961" containerName="oc" Mar 12 00:25:57 crc kubenswrapper[4948]: I0312 00:25:57.273801 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="8689cab2-961f-4987-b988-00b097f80961" containerName="oc" Mar 12 00:25:57 crc kubenswrapper[4948]: E0312 00:25:57.273879 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="154d5896-889c-432a-b3b9-38f403f4c58c" containerName="gather" Mar 12 00:25:57 crc kubenswrapper[4948]: I0312 00:25:57.273895 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="154d5896-889c-432a-b3b9-38f403f4c58c" containerName="gather" Mar 12 00:25:57 crc kubenswrapper[4948]: I0312 00:25:57.274120 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="8689cab2-961f-4987-b988-00b097f80961" containerName="oc" Mar 12 00:25:57 crc kubenswrapper[4948]: I0312 00:25:57.274141 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="154d5896-889c-432a-b3b9-38f403f4c58c" containerName="gather" Mar 12 00:25:57 crc kubenswrapper[4948]: I0312 00:25:57.274226 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="154d5896-889c-432a-b3b9-38f403f4c58c" containerName="copy" Mar 12 00:25:57 crc kubenswrapper[4948]: I0312 00:25:57.277128 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8g2vb" Mar 12 00:25:57 crc kubenswrapper[4948]: I0312 00:25:57.284381 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8g2vb"] Mar 12 00:25:57 crc kubenswrapper[4948]: I0312 00:25:57.437937 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/286d3331-0e19-4cc8-8646-58004cfb161b-utilities\") pod \"certified-operators-8g2vb\" (UID: \"286d3331-0e19-4cc8-8646-58004cfb161b\") " pod="openshift-marketplace/certified-operators-8g2vb" Mar 12 00:25:57 crc kubenswrapper[4948]: I0312 00:25:57.438063 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/286d3331-0e19-4cc8-8646-58004cfb161b-catalog-content\") pod \"certified-operators-8g2vb\" (UID: \"286d3331-0e19-4cc8-8646-58004cfb161b\") " pod="openshift-marketplace/certified-operators-8g2vb" Mar 12 00:25:57 crc kubenswrapper[4948]: I0312 00:25:57.438113 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-dcrhs\" (UniqueName: \"kubernetes.io/projected/286d3331-0e19-4cc8-8646-58004cfb161b-kube-api-access-dcrhs\") pod \"certified-operators-8g2vb\" (UID: \"286d3331-0e19-4cc8-8646-58004cfb161b\") " pod="openshift-marketplace/certified-operators-8g2vb" Mar 12 00:25:57 crc kubenswrapper[4948]: I0312 00:25:57.539518 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/286d3331-0e19-4cc8-8646-58004cfb161b-catalog-content\") pod \"certified-operators-8g2vb\" (UID: \"286d3331-0e19-4cc8-8646-58004cfb161b\") " pod="openshift-marketplace/certified-operators-8g2vb" Mar 12 00:25:57 crc kubenswrapper[4948]: I0312 00:25:57.539599 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-dcrhs\" (UniqueName: \"kubernetes.io/projected/286d3331-0e19-4cc8-8646-58004cfb161b-kube-api-access-dcrhs\") pod \"certified-operators-8g2vb\" (UID: \"286d3331-0e19-4cc8-8646-58004cfb161b\") " pod="openshift-marketplace/certified-operators-8g2vb" Mar 12 00:25:57 crc kubenswrapper[4948]: I0312 00:25:57.539633 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/286d3331-0e19-4cc8-8646-58004cfb161b-utilities\") pod \"certified-operators-8g2vb\" (UID: \"286d3331-0e19-4cc8-8646-58004cfb161b\") " pod="openshift-marketplace/certified-operators-8g2vb" Mar 12 00:25:57 crc kubenswrapper[4948]: I0312 00:25:57.540237 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/286d3331-0e19-4cc8-8646-58004cfb161b-utilities\") pod \"certified-operators-8g2vb\" (UID: \"286d3331-0e19-4cc8-8646-58004cfb161b\") " pod="openshift-marketplace/certified-operators-8g2vb" Mar 12 00:25:57 crc kubenswrapper[4948]: I0312 00:25:57.540565 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/286d3331-0e19-4cc8-8646-58004cfb161b-catalog-content\") pod \"certified-operators-8g2vb\" (UID: \"286d3331-0e19-4cc8-8646-58004cfb161b\") " pod="openshift-marketplace/certified-operators-8g2vb" Mar 12 00:25:57 crc kubenswrapper[4948]: I0312 00:25:57.564505 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-dcrhs\" (UniqueName: \"kubernetes.io/projected/286d3331-0e19-4cc8-8646-58004cfb161b-kube-api-access-dcrhs\") pod \"certified-operators-8g2vb\" (UID: \"286d3331-0e19-4cc8-8646-58004cfb161b\") " pod="openshift-marketplace/certified-operators-8g2vb" Mar 12 00:25:57 crc kubenswrapper[4948]: I0312 00:25:57.651377 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8g2vb" Mar 12 00:25:58 crc kubenswrapper[4948]: I0312 00:25:58.107811 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-marketplace/certified-operators-8g2vb"] Mar 12 00:25:58 crc kubenswrapper[4948]: I0312 00:25:58.573915 4948 generic.go:334] "Generic (PLEG): container finished" podID="286d3331-0e19-4cc8-8646-58004cfb161b" containerID="b970344a63971d325738b39d4fe915e21782841bd90d384ba8898396f1a8264e" exitCode=0 Mar 12 00:25:58 crc kubenswrapper[4948]: I0312 00:25:58.574873 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8g2vb" event={"ID":"286d3331-0e19-4cc8-8646-58004cfb161b","Type":"ContainerDied","Data":"b970344a63971d325738b39d4fe915e21782841bd90d384ba8898396f1a8264e"} Mar 12 00:25:58 crc kubenswrapper[4948]: I0312 00:25:58.574987 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8g2vb" event={"ID":"286d3331-0e19-4cc8-8646-58004cfb161b","Type":"ContainerStarted","Data":"c1d3f9591ba02a278172b5dcb3a5eef0e940f42fd592f0e6ca84277380058d0a"} Mar 12 00:25:59 crc kubenswrapper[4948]: I0312 00:25:59.584029 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8g2vb" event={"ID":"286d3331-0e19-4cc8-8646-58004cfb161b","Type":"ContainerStarted","Data":"08d27aaf12ba7a0b6ee20821c3a1ec4f253447f4197299878e20fd0a24f48c50"} Mar 12 00:26:00 crc kubenswrapper[4948]: I0312 00:26:00.150864 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29554586-zgg6j"] Mar 12 00:26:00 crc kubenswrapper[4948]: I0312 00:26:00.152659 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29554586-zgg6j" Mar 12 00:26:00 crc kubenswrapper[4948]: I0312 00:26:00.155401 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-44g5m" Mar 12 00:26:00 crc kubenswrapper[4948]: I0312 00:26:00.155699 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 00:26:00 crc kubenswrapper[4948]: I0312 00:26:00.155744 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 00:26:00 crc kubenswrapper[4948]: I0312 00:26:00.162209 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29554586-zgg6j"] Mar 12 00:26:00 crc kubenswrapper[4948]: I0312 00:26:00.278110 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-w672p\" (UniqueName: \"kubernetes.io/projected/e57c48c2-4756-4027-a9e6-4ded04d6a30a-kube-api-access-w672p\") pod \"auto-csr-approver-29554586-zgg6j\" (UID: \"e57c48c2-4756-4027-a9e6-4ded04d6a30a\") " pod="openshift-infra/auto-csr-approver-29554586-zgg6j" Mar 12 00:26:00 crc kubenswrapper[4948]: I0312 00:26:00.379327 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-w672p\" (UniqueName: \"kubernetes.io/projected/e57c48c2-4756-4027-a9e6-4ded04d6a30a-kube-api-access-w672p\") pod \"auto-csr-approver-29554586-zgg6j\" (UID: \"e57c48c2-4756-4027-a9e6-4ded04d6a30a\") " pod="openshift-infra/auto-csr-approver-29554586-zgg6j" Mar 12 00:26:00 crc kubenswrapper[4948]: I0312 00:26:00.414570 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-w672p\" (UniqueName: \"kubernetes.io/projected/e57c48c2-4756-4027-a9e6-4ded04d6a30a-kube-api-access-w672p\") pod \"auto-csr-approver-29554586-zgg6j\" (UID: \"e57c48c2-4756-4027-a9e6-4ded04d6a30a\") " pod="openshift-infra/auto-csr-approver-29554586-zgg6j" Mar 12 00:26:00 crc kubenswrapper[4948]: I0312 00:26:00.471866 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29554586-zgg6j" Mar 12 00:26:00 crc kubenswrapper[4948]: I0312 00:26:00.596389 4948 generic.go:334] "Generic (PLEG): container finished" podID="286d3331-0e19-4cc8-8646-58004cfb161b" containerID="08d27aaf12ba7a0b6ee20821c3a1ec4f253447f4197299878e20fd0a24f48c50" exitCode=0 Mar 12 00:26:00 crc kubenswrapper[4948]: I0312 00:26:00.596442 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8g2vb" event={"ID":"286d3331-0e19-4cc8-8646-58004cfb161b","Type":"ContainerDied","Data":"08d27aaf12ba7a0b6ee20821c3a1ec4f253447f4197299878e20fd0a24f48c50"} Mar 12 00:26:00 crc kubenswrapper[4948]: I0312 00:26:00.729184 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29554586-zgg6j"] Mar 12 00:26:01 crc kubenswrapper[4948]: I0312 00:26:01.607378 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8g2vb" event={"ID":"286d3331-0e19-4cc8-8646-58004cfb161b","Type":"ContainerStarted","Data":"d12eff7d711af10c28df1803909486d605f55df291e17751fe3c854bd0ea372b"} Mar 12 00:26:01 crc kubenswrapper[4948]: I0312 00:26:01.609183 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29554586-zgg6j" event={"ID":"e57c48c2-4756-4027-a9e6-4ded04d6a30a","Type":"ContainerStarted","Data":"280b0e8634cd4148427e6ae119db7056d943da8bdf5b106b8fefd5b053734162"} Mar 12 00:26:01 crc kubenswrapper[4948]: I0312 00:26:01.631122 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-marketplace/certified-operators-8g2vb" podStartSLOduration=1.991249386 podStartE2EDuration="4.631106471s" podCreationTimestamp="2026-03-12 00:25:57 +0000 UTC" firstStartedPulling="2026-03-12 00:25:58.580629236 +0000 UTC m=+1158.036232964" lastFinishedPulling="2026-03-12 00:26:01.220486311 +0000 UTC m=+1160.676090049" observedRunningTime="2026-03-12 00:26:01.628788728 +0000 UTC m=+1161.084392496" watchObservedRunningTime="2026-03-12 00:26:01.631106471 +0000 UTC m=+1161.086710209" Mar 12 00:26:02 crc kubenswrapper[4948]: I0312 00:26:02.619667 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29554586-zgg6j" event={"ID":"e57c48c2-4756-4027-a9e6-4ded04d6a30a","Type":"ContainerStarted","Data":"1d5933d6885aab553f2f3c2a3496311c8e7fd1701b70449509eb21360573e1fb"} Mar 12 00:26:02 crc kubenswrapper[4948]: I0312 00:26:02.638069 4948 pod_startup_latency_tracker.go:104] "Observed pod startup duration" pod="openshift-infra/auto-csr-approver-29554586-zgg6j" podStartSLOduration=1.181627065 podStartE2EDuration="2.638050175s" podCreationTimestamp="2026-03-12 00:26:00 +0000 UTC" firstStartedPulling="2026-03-12 00:26:00.739083529 +0000 UTC m=+1160.194687287" lastFinishedPulling="2026-03-12 00:26:02.195506629 +0000 UTC m=+1161.651110397" observedRunningTime="2026-03-12 00:26:02.632590696 +0000 UTC m=+1162.088194444" watchObservedRunningTime="2026-03-12 00:26:02.638050175 +0000 UTC m=+1162.093653943" Mar 12 00:26:03 crc kubenswrapper[4948]: I0312 00:26:03.630126 4948 generic.go:334] "Generic (PLEG): container finished" podID="e57c48c2-4756-4027-a9e6-4ded04d6a30a" containerID="1d5933d6885aab553f2f3c2a3496311c8e7fd1701b70449509eb21360573e1fb" exitCode=0 Mar 12 00:26:03 crc kubenswrapper[4948]: I0312 00:26:03.630238 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29554586-zgg6j" event={"ID":"e57c48c2-4756-4027-a9e6-4ded04d6a30a","Type":"ContainerDied","Data":"1d5933d6885aab553f2f3c2a3496311c8e7fd1701b70449509eb21360573e1fb"} Mar 12 00:26:05 crc kubenswrapper[4948]: I0312 00:26:05.005488 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29554586-zgg6j" Mar 12 00:26:05 crc kubenswrapper[4948]: I0312 00:26:05.098313 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-w672p\" (UniqueName: \"kubernetes.io/projected/e57c48c2-4756-4027-a9e6-4ded04d6a30a-kube-api-access-w672p\") pod \"e57c48c2-4756-4027-a9e6-4ded04d6a30a\" (UID: \"e57c48c2-4756-4027-a9e6-4ded04d6a30a\") " Mar 12 00:26:05 crc kubenswrapper[4948]: I0312 00:26:05.103484 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/e57c48c2-4756-4027-a9e6-4ded04d6a30a-kube-api-access-w672p" (OuterVolumeSpecName: "kube-api-access-w672p") pod "e57c48c2-4756-4027-a9e6-4ded04d6a30a" (UID: "e57c48c2-4756-4027-a9e6-4ded04d6a30a"). InnerVolumeSpecName "kube-api-access-w672p". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:26:05 crc kubenswrapper[4948]: I0312 00:26:05.199897 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-w672p\" (UniqueName: \"kubernetes.io/projected/e57c48c2-4756-4027-a9e6-4ded04d6a30a-kube-api-access-w672p\") on node \"crc\" DevicePath \"\"" Mar 12 00:26:05 crc kubenswrapper[4948]: I0312 00:26:05.650834 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29554586-zgg6j" event={"ID":"e57c48c2-4756-4027-a9e6-4ded04d6a30a","Type":"ContainerDied","Data":"280b0e8634cd4148427e6ae119db7056d943da8bdf5b106b8fefd5b053734162"} Mar 12 00:26:05 crc kubenswrapper[4948]: I0312 00:26:05.650910 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="280b0e8634cd4148427e6ae119db7056d943da8bdf5b106b8fefd5b053734162" Mar 12 00:26:05 crc kubenswrapper[4948]: I0312 00:26:05.651015 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29554586-zgg6j" Mar 12 00:26:05 crc kubenswrapper[4948]: I0312 00:26:05.717637 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29554580-ghvkh"] Mar 12 00:26:05 crc kubenswrapper[4948]: I0312 00:26:05.723661 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29554580-ghvkh"] Mar 12 00:26:07 crc kubenswrapper[4948]: I0312 00:26:07.326476 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="4696b53e-388b-457f-9dc9-93811078ccc4" path="/var/lib/kubelet/pods/4696b53e-388b-457f-9dc9-93811078ccc4/volumes" Mar 12 00:26:07 crc kubenswrapper[4948]: I0312 00:26:07.651871 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="unhealthy" pod="openshift-marketplace/certified-operators-8g2vb" Mar 12 00:26:07 crc kubenswrapper[4948]: I0312 00:26:07.651941 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="" pod="openshift-marketplace/certified-operators-8g2vb" Mar 12 00:26:07 crc kubenswrapper[4948]: I0312 00:26:07.714570 4948 kubelet.go:2542] "SyncLoop (probe)" probe="startup" status="started" pod="openshift-marketplace/certified-operators-8g2vb" Mar 12 00:26:07 crc kubenswrapper[4948]: I0312 00:26:07.787423 4948 kubelet.go:2542] "SyncLoop (probe)" probe="readiness" status="ready" pod="openshift-marketplace/certified-operators-8g2vb" Mar 12 00:26:07 crc kubenswrapper[4948]: I0312 00:26:07.963946 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8g2vb"] Mar 12 00:26:09 crc kubenswrapper[4948]: I0312 00:26:09.682337 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-marketplace/certified-operators-8g2vb" podUID="286d3331-0e19-4cc8-8646-58004cfb161b" containerName="registry-server" containerID="cri-o://d12eff7d711af10c28df1803909486d605f55df291e17751fe3c854bd0ea372b" gracePeriod=2 Mar 12 00:26:10 crc kubenswrapper[4948]: I0312 00:26:10.128547 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8g2vb" Mar 12 00:26:10 crc kubenswrapper[4948]: I0312 00:26:10.269882 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/286d3331-0e19-4cc8-8646-58004cfb161b-utilities\") pod \"286d3331-0e19-4cc8-8646-58004cfb161b\" (UID: \"286d3331-0e19-4cc8-8646-58004cfb161b\") " Mar 12 00:26:10 crc kubenswrapper[4948]: I0312 00:26:10.269946 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-dcrhs\" (UniqueName: \"kubernetes.io/projected/286d3331-0e19-4cc8-8646-58004cfb161b-kube-api-access-dcrhs\") pod \"286d3331-0e19-4cc8-8646-58004cfb161b\" (UID: \"286d3331-0e19-4cc8-8646-58004cfb161b\") " Mar 12 00:26:10 crc kubenswrapper[4948]: I0312 00:26:10.270032 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/286d3331-0e19-4cc8-8646-58004cfb161b-catalog-content\") pod \"286d3331-0e19-4cc8-8646-58004cfb161b\" (UID: \"286d3331-0e19-4cc8-8646-58004cfb161b\") " Mar 12 00:26:10 crc kubenswrapper[4948]: I0312 00:26:10.271742 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/286d3331-0e19-4cc8-8646-58004cfb161b-utilities" (OuterVolumeSpecName: "utilities") pod "286d3331-0e19-4cc8-8646-58004cfb161b" (UID: "286d3331-0e19-4cc8-8646-58004cfb161b"). InnerVolumeSpecName "utilities". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 00:26:10 crc kubenswrapper[4948]: I0312 00:26:10.279326 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/286d3331-0e19-4cc8-8646-58004cfb161b-kube-api-access-dcrhs" (OuterVolumeSpecName: "kube-api-access-dcrhs") pod "286d3331-0e19-4cc8-8646-58004cfb161b" (UID: "286d3331-0e19-4cc8-8646-58004cfb161b"). InnerVolumeSpecName "kube-api-access-dcrhs". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:26:10 crc kubenswrapper[4948]: I0312 00:26:10.372269 4948 reconciler_common.go:293] "Volume detached for volume \"utilities\" (UniqueName: \"kubernetes.io/empty-dir/286d3331-0e19-4cc8-8646-58004cfb161b-utilities\") on node \"crc\" DevicePath \"\"" Mar 12 00:26:10 crc kubenswrapper[4948]: I0312 00:26:10.372356 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-dcrhs\" (UniqueName: \"kubernetes.io/projected/286d3331-0e19-4cc8-8646-58004cfb161b-kube-api-access-dcrhs\") on node \"crc\" DevicePath \"\"" Mar 12 00:26:10 crc kubenswrapper[4948]: I0312 00:26:10.697704 4948 generic.go:334] "Generic (PLEG): container finished" podID="286d3331-0e19-4cc8-8646-58004cfb161b" containerID="d12eff7d711af10c28df1803909486d605f55df291e17751fe3c854bd0ea372b" exitCode=0 Mar 12 00:26:10 crc kubenswrapper[4948]: I0312 00:26:10.697757 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8g2vb" event={"ID":"286d3331-0e19-4cc8-8646-58004cfb161b","Type":"ContainerDied","Data":"d12eff7d711af10c28df1803909486d605f55df291e17751fe3c854bd0ea372b"} Mar 12 00:26:10 crc kubenswrapper[4948]: I0312 00:26:10.697808 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-marketplace/certified-operators-8g2vb" event={"ID":"286d3331-0e19-4cc8-8646-58004cfb161b","Type":"ContainerDied","Data":"c1d3f9591ba02a278172b5dcb3a5eef0e940f42fd592f0e6ca84277380058d0a"} Mar 12 00:26:10 crc kubenswrapper[4948]: I0312 00:26:10.697807 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-marketplace/certified-operators-8g2vb" Mar 12 00:26:10 crc kubenswrapper[4948]: I0312 00:26:10.697923 4948 scope.go:117] "RemoveContainer" containerID="d12eff7d711af10c28df1803909486d605f55df291e17751fe3c854bd0ea372b" Mar 12 00:26:10 crc kubenswrapper[4948]: I0312 00:26:10.724902 4948 scope.go:117] "RemoveContainer" containerID="08d27aaf12ba7a0b6ee20821c3a1ec4f253447f4197299878e20fd0a24f48c50" Mar 12 00:26:10 crc kubenswrapper[4948]: I0312 00:26:10.750916 4948 scope.go:117] "RemoveContainer" containerID="b970344a63971d325738b39d4fe915e21782841bd90d384ba8898396f1a8264e" Mar 12 00:26:10 crc kubenswrapper[4948]: I0312 00:26:10.793666 4948 scope.go:117] "RemoveContainer" containerID="d12eff7d711af10c28df1803909486d605f55df291e17751fe3c854bd0ea372b" Mar 12 00:26:10 crc kubenswrapper[4948]: E0312 00:26:10.794536 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"d12eff7d711af10c28df1803909486d605f55df291e17751fe3c854bd0ea372b\": container with ID starting with d12eff7d711af10c28df1803909486d605f55df291e17751fe3c854bd0ea372b not found: ID does not exist" containerID="d12eff7d711af10c28df1803909486d605f55df291e17751fe3c854bd0ea372b" Mar 12 00:26:10 crc kubenswrapper[4948]: I0312 00:26:10.794603 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"d12eff7d711af10c28df1803909486d605f55df291e17751fe3c854bd0ea372b"} err="failed to get container status \"d12eff7d711af10c28df1803909486d605f55df291e17751fe3c854bd0ea372b\": rpc error: code = NotFound desc = could not find container \"d12eff7d711af10c28df1803909486d605f55df291e17751fe3c854bd0ea372b\": container with ID starting with d12eff7d711af10c28df1803909486d605f55df291e17751fe3c854bd0ea372b not found: ID does not exist" Mar 12 00:26:10 crc kubenswrapper[4948]: I0312 00:26:10.794645 4948 scope.go:117] "RemoveContainer" containerID="08d27aaf12ba7a0b6ee20821c3a1ec4f253447f4197299878e20fd0a24f48c50" Mar 12 00:26:10 crc kubenswrapper[4948]: E0312 00:26:10.795425 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"08d27aaf12ba7a0b6ee20821c3a1ec4f253447f4197299878e20fd0a24f48c50\": container with ID starting with 08d27aaf12ba7a0b6ee20821c3a1ec4f253447f4197299878e20fd0a24f48c50 not found: ID does not exist" containerID="08d27aaf12ba7a0b6ee20821c3a1ec4f253447f4197299878e20fd0a24f48c50" Mar 12 00:26:10 crc kubenswrapper[4948]: I0312 00:26:10.795475 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"08d27aaf12ba7a0b6ee20821c3a1ec4f253447f4197299878e20fd0a24f48c50"} err="failed to get container status \"08d27aaf12ba7a0b6ee20821c3a1ec4f253447f4197299878e20fd0a24f48c50\": rpc error: code = NotFound desc = could not find container \"08d27aaf12ba7a0b6ee20821c3a1ec4f253447f4197299878e20fd0a24f48c50\": container with ID starting with 08d27aaf12ba7a0b6ee20821c3a1ec4f253447f4197299878e20fd0a24f48c50 not found: ID does not exist" Mar 12 00:26:10 crc kubenswrapper[4948]: I0312 00:26:10.795503 4948 scope.go:117] "RemoveContainer" containerID="b970344a63971d325738b39d4fe915e21782841bd90d384ba8898396f1a8264e" Mar 12 00:26:10 crc kubenswrapper[4948]: E0312 00:26:10.796108 4948 log.go:32] "ContainerStatus from runtime service failed" err="rpc error: code = NotFound desc = could not find container \"b970344a63971d325738b39d4fe915e21782841bd90d384ba8898396f1a8264e\": container with ID starting with b970344a63971d325738b39d4fe915e21782841bd90d384ba8898396f1a8264e not found: ID does not exist" containerID="b970344a63971d325738b39d4fe915e21782841bd90d384ba8898396f1a8264e" Mar 12 00:26:10 crc kubenswrapper[4948]: I0312 00:26:10.796154 4948 pod_container_deletor.go:53] "DeleteContainer returned error" containerID={"Type":"cri-o","ID":"b970344a63971d325738b39d4fe915e21782841bd90d384ba8898396f1a8264e"} err="failed to get container status \"b970344a63971d325738b39d4fe915e21782841bd90d384ba8898396f1a8264e\": rpc error: code = NotFound desc = could not find container \"b970344a63971d325738b39d4fe915e21782841bd90d384ba8898396f1a8264e\": container with ID starting with b970344a63971d325738b39d4fe915e21782841bd90d384ba8898396f1a8264e not found: ID does not exist" Mar 12 00:26:10 crc kubenswrapper[4948]: I0312 00:26:10.891767 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/empty-dir/286d3331-0e19-4cc8-8646-58004cfb161b-catalog-content" (OuterVolumeSpecName: "catalog-content") pod "286d3331-0e19-4cc8-8646-58004cfb161b" (UID: "286d3331-0e19-4cc8-8646-58004cfb161b"). InnerVolumeSpecName "catalog-content". PluginName "kubernetes.io/empty-dir", VolumeGidValue "" Mar 12 00:26:10 crc kubenswrapper[4948]: I0312 00:26:10.982432 4948 reconciler_common.go:293] "Volume detached for volume \"catalog-content\" (UniqueName: \"kubernetes.io/empty-dir/286d3331-0e19-4cc8-8646-58004cfb161b-catalog-content\") on node \"crc\" DevicePath \"\"" Mar 12 00:26:11 crc kubenswrapper[4948]: I0312 00:26:11.050827 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-marketplace/certified-operators-8g2vb"] Mar 12 00:26:11 crc kubenswrapper[4948]: I0312 00:26:11.066273 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-marketplace/certified-operators-8g2vb"] Mar 12 00:26:11 crc kubenswrapper[4948]: I0312 00:26:11.326427 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="286d3331-0e19-4cc8-8646-58004cfb161b" path="/var/lib/kubelet/pods/286d3331-0e19-4cc8-8646-58004cfb161b/volumes" Mar 12 00:26:57 crc kubenswrapper[4948]: I0312 00:26:57.242692 4948 scope.go:117] "RemoveContainer" containerID="381301100f845275e861e7f705307b29cde25017428f668242d0e84e87383c4a" Mar 12 00:27:44 crc kubenswrapper[4948]: I0312 00:27:44.785246 4948 patch_prober.go:28] interesting pod/machine-config-daemon-m4xwc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 00:27:44 crc kubenswrapper[4948]: I0312 00:27:44.788195 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" podUID="7e483c2b-08f0-4e92-8e4a-b7281f30af3e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 00:28:00 crc kubenswrapper[4948]: I0312 00:28:00.159207 4948 kubelet.go:2421] "SyncLoop ADD" source="api" pods=["openshift-infra/auto-csr-approver-29554588-kvn5g"] Mar 12 00:28:00 crc kubenswrapper[4948]: E0312 00:28:00.160160 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="286d3331-0e19-4cc8-8646-58004cfb161b" containerName="extract-content" Mar 12 00:28:00 crc kubenswrapper[4948]: I0312 00:28:00.160175 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="286d3331-0e19-4cc8-8646-58004cfb161b" containerName="extract-content" Mar 12 00:28:00 crc kubenswrapper[4948]: E0312 00:28:00.160193 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="e57c48c2-4756-4027-a9e6-4ded04d6a30a" containerName="oc" Mar 12 00:28:00 crc kubenswrapper[4948]: I0312 00:28:00.160200 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="e57c48c2-4756-4027-a9e6-4ded04d6a30a" containerName="oc" Mar 12 00:28:00 crc kubenswrapper[4948]: E0312 00:28:00.160219 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="286d3331-0e19-4cc8-8646-58004cfb161b" containerName="registry-server" Mar 12 00:28:00 crc kubenswrapper[4948]: I0312 00:28:00.160226 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="286d3331-0e19-4cc8-8646-58004cfb161b" containerName="registry-server" Mar 12 00:28:00 crc kubenswrapper[4948]: E0312 00:28:00.160236 4948 cpu_manager.go:410] "RemoveStaleState: removing container" podUID="286d3331-0e19-4cc8-8646-58004cfb161b" containerName="extract-utilities" Mar 12 00:28:00 crc kubenswrapper[4948]: I0312 00:28:00.160242 4948 state_mem.go:107] "Deleted CPUSet assignment" podUID="286d3331-0e19-4cc8-8646-58004cfb161b" containerName="extract-utilities" Mar 12 00:28:00 crc kubenswrapper[4948]: I0312 00:28:00.160426 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="e57c48c2-4756-4027-a9e6-4ded04d6a30a" containerName="oc" Mar 12 00:28:00 crc kubenswrapper[4948]: I0312 00:28:00.160451 4948 memory_manager.go:354] "RemoveStaleState removing state" podUID="286d3331-0e19-4cc8-8646-58004cfb161b" containerName="registry-server" Mar 12 00:28:00 crc kubenswrapper[4948]: I0312 00:28:00.162070 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29554588-kvn5g" Mar 12 00:28:00 crc kubenswrapper[4948]: I0312 00:28:00.169111 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"openshift-service-ca.crt" Mar 12 00:28:00 crc kubenswrapper[4948]: I0312 00:28:00.169105 4948 reflector.go:368] Caches populated for *v1.Secret from object-"openshift-infra"/"csr-approver-sa-dockercfg-44g5m" Mar 12 00:28:00 crc kubenswrapper[4948]: I0312 00:28:00.169686 4948 reflector.go:368] Caches populated for *v1.ConfigMap from object-"openshift-infra"/"kube-root-ca.crt" Mar 12 00:28:00 crc kubenswrapper[4948]: I0312 00:28:00.171822 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29554588-kvn5g"] Mar 12 00:28:00 crc kubenswrapper[4948]: I0312 00:28:00.275037 4948 reconciler_common.go:245] "operationExecutor.VerifyControllerAttachedVolume started for volume \"kube-api-access-gqgsg\" (UniqueName: \"kubernetes.io/projected/25bf4a0a-b2ca-46ff-a0f7-833ae485bbf4-kube-api-access-gqgsg\") pod \"auto-csr-approver-29554588-kvn5g\" (UID: \"25bf4a0a-b2ca-46ff-a0f7-833ae485bbf4\") " pod="openshift-infra/auto-csr-approver-29554588-kvn5g" Mar 12 00:28:00 crc kubenswrapper[4948]: I0312 00:28:00.376961 4948 reconciler_common.go:218] "operationExecutor.MountVolume started for volume \"kube-api-access-gqgsg\" (UniqueName: \"kubernetes.io/projected/25bf4a0a-b2ca-46ff-a0f7-833ae485bbf4-kube-api-access-gqgsg\") pod \"auto-csr-approver-29554588-kvn5g\" (UID: \"25bf4a0a-b2ca-46ff-a0f7-833ae485bbf4\") " pod="openshift-infra/auto-csr-approver-29554588-kvn5g" Mar 12 00:28:00 crc kubenswrapper[4948]: I0312 00:28:00.406087 4948 operation_generator.go:637] "MountVolume.SetUp succeeded for volume \"kube-api-access-gqgsg\" (UniqueName: \"kubernetes.io/projected/25bf4a0a-b2ca-46ff-a0f7-833ae485bbf4-kube-api-access-gqgsg\") pod \"auto-csr-approver-29554588-kvn5g\" (UID: \"25bf4a0a-b2ca-46ff-a0f7-833ae485bbf4\") " pod="openshift-infra/auto-csr-approver-29554588-kvn5g" Mar 12 00:28:00 crc kubenswrapper[4948]: I0312 00:28:00.488940 4948 util.go:30] "No sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29554588-kvn5g" Mar 12 00:28:00 crc kubenswrapper[4948]: I0312 00:28:00.752475 4948 kubelet.go:2428] "SyncLoop UPDATE" source="api" pods=["openshift-infra/auto-csr-approver-29554588-kvn5g"] Mar 12 00:28:00 crc kubenswrapper[4948]: W0312 00:28:00.775646 4948 manager.go:1169] Failed to process watch event {EventType:0 Name:/kubepods.slice/kubepods-besteffort.slice/kubepods-besteffort-pod25bf4a0a_b2ca_46ff_a0f7_833ae485bbf4.slice/crio-dfe675e4aa9c43513bcc594edae65c7ffa4fc651cfbffdd7c0315dc2da7ce7cb WatchSource:0}: Error finding container dfe675e4aa9c43513bcc594edae65c7ffa4fc651cfbffdd7c0315dc2da7ce7cb: Status 404 returned error can't find the container with id dfe675e4aa9c43513bcc594edae65c7ffa4fc651cfbffdd7c0315dc2da7ce7cb Mar 12 00:28:00 crc kubenswrapper[4948]: I0312 00:28:00.778182 4948 provider.go:102] Refreshing cache for provider: *credentialprovider.defaultDockerConfigProvider Mar 12 00:28:01 crc kubenswrapper[4948]: I0312 00:28:01.577489 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29554588-kvn5g" event={"ID":"25bf4a0a-b2ca-46ff-a0f7-833ae485bbf4","Type":"ContainerStarted","Data":"dfe675e4aa9c43513bcc594edae65c7ffa4fc651cfbffdd7c0315dc2da7ce7cb"} Mar 12 00:28:02 crc kubenswrapper[4948]: I0312 00:28:02.584553 4948 generic.go:334] "Generic (PLEG): container finished" podID="25bf4a0a-b2ca-46ff-a0f7-833ae485bbf4" containerID="87cb2d0e478e2fd3ce669f044715d9b8d0619c435053c29c3e7e90650aec95ff" exitCode=0 Mar 12 00:28:02 crc kubenswrapper[4948]: I0312 00:28:02.584595 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29554588-kvn5g" event={"ID":"25bf4a0a-b2ca-46ff-a0f7-833ae485bbf4","Type":"ContainerDied","Data":"87cb2d0e478e2fd3ce669f044715d9b8d0619c435053c29c3e7e90650aec95ff"} Mar 12 00:28:03 crc kubenswrapper[4948]: I0312 00:28:03.930332 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29554588-kvn5g" Mar 12 00:28:04 crc kubenswrapper[4948]: I0312 00:28:04.027363 4948 reconciler_common.go:159] "operationExecutor.UnmountVolume started for volume \"kube-api-access-gqgsg\" (UniqueName: \"kubernetes.io/projected/25bf4a0a-b2ca-46ff-a0f7-833ae485bbf4-kube-api-access-gqgsg\") pod \"25bf4a0a-b2ca-46ff-a0f7-833ae485bbf4\" (UID: \"25bf4a0a-b2ca-46ff-a0f7-833ae485bbf4\") " Mar 12 00:28:04 crc kubenswrapper[4948]: I0312 00:28:04.032587 4948 operation_generator.go:803] UnmountVolume.TearDown succeeded for volume "kubernetes.io/projected/25bf4a0a-b2ca-46ff-a0f7-833ae485bbf4-kube-api-access-gqgsg" (OuterVolumeSpecName: "kube-api-access-gqgsg") pod "25bf4a0a-b2ca-46ff-a0f7-833ae485bbf4" (UID: "25bf4a0a-b2ca-46ff-a0f7-833ae485bbf4"). InnerVolumeSpecName "kube-api-access-gqgsg". PluginName "kubernetes.io/projected", VolumeGidValue "" Mar 12 00:28:04 crc kubenswrapper[4948]: I0312 00:28:04.129787 4948 reconciler_common.go:293] "Volume detached for volume \"kube-api-access-gqgsg\" (UniqueName: \"kubernetes.io/projected/25bf4a0a-b2ca-46ff-a0f7-833ae485bbf4-kube-api-access-gqgsg\") on node \"crc\" DevicePath \"\"" Mar 12 00:28:04 crc kubenswrapper[4948]: I0312 00:28:04.605449 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-infra/auto-csr-approver-29554588-kvn5g" event={"ID":"25bf4a0a-b2ca-46ff-a0f7-833ae485bbf4","Type":"ContainerDied","Data":"dfe675e4aa9c43513bcc594edae65c7ffa4fc651cfbffdd7c0315dc2da7ce7cb"} Mar 12 00:28:04 crc kubenswrapper[4948]: I0312 00:28:04.605521 4948 pod_container_deletor.go:80] "Container not found in pod's containers" containerID="dfe675e4aa9c43513bcc594edae65c7ffa4fc651cfbffdd7c0315dc2da7ce7cb" Mar 12 00:28:04 crc kubenswrapper[4948]: I0312 00:28:04.605633 4948 util.go:48] "No ready sandbox for pod can be found. Need to start a new one" pod="openshift-infra/auto-csr-approver-29554588-kvn5g" Mar 12 00:28:05 crc kubenswrapper[4948]: I0312 00:28:05.016455 4948 kubelet.go:2437] "SyncLoop DELETE" source="api" pods=["openshift-infra/auto-csr-approver-29554582-4h2q8"] Mar 12 00:28:05 crc kubenswrapper[4948]: I0312 00:28:05.025471 4948 kubelet.go:2431] "SyncLoop REMOVE" source="api" pods=["openshift-infra/auto-csr-approver-29554582-4h2q8"] Mar 12 00:28:05 crc kubenswrapper[4948]: I0312 00:28:05.327895 4948 kubelet_volumes.go:163] "Cleaned up orphaned pod volumes dir" podUID="98e5bf0e-b218-4f0c-a23f-afdb5ef99139" path="/var/lib/kubelet/pods/98e5bf0e-b218-4f0c-a23f-afdb5ef99139/volumes" Mar 12 00:28:14 crc kubenswrapper[4948]: I0312 00:28:14.784875 4948 patch_prober.go:28] interesting pod/machine-config-daemon-m4xwc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 00:28:14 crc kubenswrapper[4948]: I0312 00:28:14.785483 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" podUID="7e483c2b-08f0-4e92-8e4a-b7281f30af3e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 00:28:44 crc kubenswrapper[4948]: I0312 00:28:44.785041 4948 patch_prober.go:28] interesting pod/machine-config-daemon-m4xwc container/machine-config-daemon namespace/openshift-machine-config-operator: Liveness probe status=failure output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" start-of-body= Mar 12 00:28:44 crc kubenswrapper[4948]: I0312 00:28:44.785543 4948 prober.go:107] "Probe failed" probeType="Liveness" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" podUID="7e483c2b-08f0-4e92-8e4a-b7281f30af3e" containerName="machine-config-daemon" probeResult="failure" output="Get \"http://127.0.0.1:8798/health\": dial tcp 127.0.0.1:8798: connect: connection refused" Mar 12 00:28:44 crc kubenswrapper[4948]: I0312 00:28:44.785603 4948 kubelet.go:2542] "SyncLoop (probe)" probe="liveness" status="unhealthy" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" Mar 12 00:28:44 crc kubenswrapper[4948]: I0312 00:28:44.786539 4948 kuberuntime_manager.go:1027] "Message for Container of pod" containerName="machine-config-daemon" containerStatusID={"Type":"cri-o","ID":"4207332ab030bc6bdba00683e8604b734b37d7a5dcb4073592c80c1f073959b8"} pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" containerMessage="Container machine-config-daemon failed liveness probe, will be restarted" Mar 12 00:28:44 crc kubenswrapper[4948]: I0312 00:28:44.786655 4948 kuberuntime_container.go:808] "Killing container with a grace period" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" podUID="7e483c2b-08f0-4e92-8e4a-b7281f30af3e" containerName="machine-config-daemon" containerID="cri-o://4207332ab030bc6bdba00683e8604b734b37d7a5dcb4073592c80c1f073959b8" gracePeriod=600 Mar 12 00:28:44 crc kubenswrapper[4948]: I0312 00:28:44.985084 4948 generic.go:334] "Generic (PLEG): container finished" podID="7e483c2b-08f0-4e92-8e4a-b7281f30af3e" containerID="4207332ab030bc6bdba00683e8604b734b37d7a5dcb4073592c80c1f073959b8" exitCode=0 Mar 12 00:28:44 crc kubenswrapper[4948]: I0312 00:28:44.985146 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" event={"ID":"7e483c2b-08f0-4e92-8e4a-b7281f30af3e","Type":"ContainerDied","Data":"4207332ab030bc6bdba00683e8604b734b37d7a5dcb4073592c80c1f073959b8"} Mar 12 00:28:44 crc kubenswrapper[4948]: I0312 00:28:44.985218 4948 scope.go:117] "RemoveContainer" containerID="8f5f07d17ca670886abddff8cc1ea5d57cf97a7dc2b3744fe1219d1f6e701f88" Mar 12 00:28:46 crc kubenswrapper[4948]: I0312 00:28:46.007565 4948 kubelet.go:2453] "SyncLoop (PLEG): event for pod" pod="openshift-machine-config-operator/machine-config-daemon-m4xwc" event={"ID":"7e483c2b-08f0-4e92-8e4a-b7281f30af3e","Type":"ContainerStarted","Data":"5fad5b4fb905e753ed933046087a980a62760efd434b9e0c9f19a2515690764d"} Mar 12 00:28:57 crc kubenswrapper[4948]: I0312 00:28:57.339933 4948 scope.go:117] "RemoveContainer" containerID="78ae7ff1e7adb8335eb10046583835dd6e974d34218d47686d62115ff0197d49" var/home/core/zuul-output/logs/crc-cloud-workdir-crc-all-logs.tar.gz0000644000175000000000000000005515154404354024452 0ustar coreroot  Om77'(var/home/core/zuul-output/logs/crc-cloud/0000755000175000000000000000000015154404354017367 5ustar corerootvar/home/core/zuul-output/artifacts/0000755000175000017500000000000015154401252016504 5ustar corecorevar/home/core/zuul-output/docs/0000755000175000017500000000000015154401252015454 5ustar corecore